var/home/core/zuul-output/0000755000175000017500000000000015136646002014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015136655301015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000242134615136655127020275 0ustar corecoreWZ{ikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD  >KEڤ펯_ˎ6Ϸ7+%f?長ox[o8W5Ι!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁weor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/P_]F@?qr7@sON_}ۿ릶ytoyמseQv^sP3.sP1'Ns}d_ս=f1Jid % Jwe`40^|ǜd]z dJR-Дxq4lZ,Z[|e 'Ƙ$b2JOh k[b>¾h[;:>OM=y)֖[Sm5*_?$cjf `~ߛUIOvl/.4`P{d056 %w ^?sʫ"nK)D}O >%9r}1j#e[tRQ9*ء !ǨLJ- upƜ/4cY\[|Xs;ɾ7-<S1wg y &SL9qk;NP> ,wդjtah-j:_[;4Wg_0K>є0vNۈ/ze={< 1;/STcD,ڙ`[3XPo0TXx ZYޏ=S-ܑ2Uڞ7կZ8m1`qAewQT*:ÊxtŨ!u}$K6tem@t):êtx: `)L`m GƂ%k1羨(zv:U!2`cV, lNdV5m$/KFS#0gLwNO6¨h}'XvوPkWn}/7d*1q* c0.$\+XND]P*84[߷Q뽃J޸8iD WPC49 *#LC ءzCwS%'m'3ܚ|otoʉ!9:PZ"ρ5M^kVځIX%G^{;+Fi7Z(ZN~;MM/u2}ݼPݫedKAd#[ BeMP6" YǨ 0vyv?7R F"}8&q]ows!Z!C4g*8n]rMQ ;N>Sr??Ӽ]\+hSQזL +3[n )ܗKj/jUSsȕD $([LH%xa1yrO('h=TԫeVިO? )-1p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*ؠqi\9#.\r= (mO(f=rWmd'rDZ~;o\mkmB`s ~7!GdјCyEߖs|n|zu0VhI|/{}BC6q>HĜ]Xgy G[Ŷ.|37xo=N4wjDH>:&EOΆ<䧊1v@b&툒f!yO){~%gq~.LK78F#E01g.u7^Ew_lv۠M0}qk:Lx%` urJp)>I(>z`{|puB"8#YkrZ .`h(eek[?̱ՒOOc&!dVzMEHH*V"MC Qؽ1Omsz/v0vȌJBIG,CNˆ-L{L #cNqgVR2r뭲⭊ڰ08uirP qNUӛ<|߈$m뫷dùB Z^-_dsz=F8jH˽&DUh+9k̈́W^̤F˖.kL5̻wS"!5<@&] WE\wMc%={_bD&k 5:lb69OBCC*F!6~ö9M( Pnuݮ)`Q6eMӁKzFZf;5IW1i[xU 0FPM]gl}>6sUDO5f p6mD[%ZZvm̓'!n&.TU n$%rIwP(fwnv :Nb=X~ax`;Vw}wvRS1q!z989ep 5w%ZU.]5`s=r&v2FaUM 6/"IiBSpp3n_9>Byݝ0_5bZ8ւ 6{Sf觋-V=Oߖm!6jm3Kx6BDhvzZn8hSlz z6^Q1* _> 8A@>!a:dC<mWu[7-D[9)/*˸PP!j-7BtK|VXnT&eZc~=31mס̈'K^r,W˲vtv|,SԽ[qɑ)6&vד4G&%JLi[? 1A ۥ͟յt9 ",@9 P==s 0py(nWDwpɡ`i?E1Q!:5*6@q\\YWTk sspww0SZ2, uvao=\Sl Uݚu@$Pup՗з҃TXskwqRtYڢLhw KO5C\-&-qQ4Mv8pS俺kCߤ`ZnTV*P,rq<-mOK[[ߢm۽ȑt^, tJbظ&Pg%㢒\QS܁vn` *3UP0Sp8:>m(Zx ,c|!0=0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0<=)W}Q\lZcmONii@gIi3]v_ r[Lo{o# yTqM8@4Q2lE >)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==qgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|\΋"?|NKfֱn !-p^:ZYUv`Ƌ-v|u>r,8.7uO`c Nc0%Ն R C%_ EV a"҅4 |T!DdǍ- .™5,V:;[g./0 +v䤗dWF >:֓[@ QPltsHtQ$J==O!;*>ohǖVa[|E7e0ϕ9Uyzg%pg/cc6RS`HFLЩ LkJu\!`0);Sak$Vfp~C%YdE6c>1ƕ (0W4Q>@>lWN"^ X5G-nm.8B>NOI[31,j2 Ce |M>8l WIf|\q4|UkC.gr`˱Lϰ} xr.~l-ɩu_Drd31V_ѺUib0/ %IYhq ҕ  O UA!wY~ -`%Űb`\mS38W1`vOF7/.C!Pu&Jm l?Q>}O+D7 P=x@`0ʿ26a>d Bqε^a'NԋsI`Yu.7v$Rt)Ag:ݙyX|HkX cU82IP qgzkX=>׻K߉J%E92' ]qҙ%rXgs+"sc9| ]>T]"JرWBΌ-zJS-~y30G@U#=h7) ^EUB Q:>9W΀çM{?`c`uRljצXr:l`T~IQg\Ѝpgu#QH! ,/3`~eB|C1Yg~ؼ/5I7w9I}qww}U~7뭱ԏ,}e7]ukDn`jSlQ7DžHa/EU^IpYWW兹Q7WyTz|nˇ _qˍ[!;n ^b k[);ng]ȶM_u)O_xV hx h[K2kـ`b duhq[..cS'5YO@˒ӓdcY'HAKq^$8`b $1r Qz?ۧ1ZM/G+qYcYl YhD$kt_TId E$dS:֢̆ ?GЅ'JƖ'ZXO݇'kJՂU086\h%1GK(Yn% ']Q; Gd:!gI-XEmkF}:~0}4t3Qf5xd\hEB-} |q*ȃThLj'sQ %؇Gk`F;Sl\h)5؈x2Ld="KԦ:EVewN ًS9d#$*u>>I#lX9vW !&H2kVyKZt<cm^] bCD6b&>9VE7e4p +{&g߷2KY,`Wf1_ܑMYٚ'`ySc4ΔV`nI+ƳC6;җ2ct"*5S}t)eNqǪP@o`co ˎ<عLۀG\ 7۶+q|YRiĹ zm/bcK3;=,7}RqT vvFI O0]&5uKMf#pDTk6yi*cem:y0W|1u CWL;oG^\ X5.aRߦ[_Vs? Ž^A12JQ̛XL:OEUپOY>WK-uP0\8"M: /P4Qz~j3 .-8NJ|!N9/|a|>lX9T ҇t~T1=UF"t; 8-1I|2L+)WȱL˿ˍ-038D*0-)ZyT13`tTnm|Yhi+lQ&Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ6UNE@Uo/>8.fgW]kY0Cgcu6/!_Ɩ} ' Ў3)X<seWfSv!ؒRKfs%(1Lhrٵ L.] s?I,HCԢ[b C-lLG+@_$c%* _jR|\:dc5u= A@kUc\ǔz;M>dUN/aFRĦ@x؂ǀ$6%}N^ \mQ!%8j0dUo=rh>*YȴU3Q,̸*E%59sTzɟڮ2kg ۱wEUD3uKrr&"B:p`\E)j<).R&#ÃecE,dp"nPS 44 Q8ZƈKnnJei+^z '3JDbSK;*uБ:hF ѹ @˿ޗ~7g9| hLXULi7.1-Qk%ƩJ4^=ple;u.6vQe UZAl *^Vif]>HUd6ƕ̽=T/se+ϙK$S`hnOcE(Tcr!:8UL | 8 !t Q7jk=nn7J0ܽ0{GGL'_So^ʮL_'s%eU+U+ȳlX6}i@djӃfb -u-w~ r}plK;ֽ=nlmuo[`wdй d:[mS%uTڪ?>={2])|Ը>U{s]^l`+ ja^9c5~nZjA|ЩJs Va[~ۗ#rri# zLdMl?6o AMҪ1Ez&I2Wwߎ|7.sW\zk﯊溺^TW^T\*6eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th $LqQjiXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1M'I`22؆DT!/j璓P åiw@wgRCsT~$U>ceއE)BI>UljO|Ty$ŋrwOtZ7$ "i 8U 7bSem'k?I+/ScYƝW|/-R%pm`6m"$EU=xJcIx֍&sřIRգ$n̓*'[rP&($8Tš g{|MA"Rn8T(ET<#ʢL*Yvh±7UP*F`rEH`#\8Xy4eUq{_~ σL%6qP |MA ,y@`5Ń<8TMy }oqqp #PEaLyQz0(h(=jr/U #O<™8:srxO AE5=x>L&DҞ䑸9\3}\$7I9Nr :ϐ=2,m{<%e+BtJV?:u "OBBFϙGIȞ]Ke/B3AUqcQb~nN8 UC[yQ1%~gm((#VPշ_J}}|W$)DQȟG'1fwrk e7 q8"|wL3^LaCCިKMDF!)g}sP>IK^Rdi9c$ X k&ò>hFV̀D$8ߴ4>mʨ2r1нYz x}eQ!>qx?)L:C+:(C_r[ #HC$Yم“{ʅ`A5ʟC-N,ySXj4,P2JXQ߫Ki(DHD&ar]zXJWކSyRx\j(e(U?3×2{g9&!gXDI-vXDlqcGpcϧqH|:h'TƊ4M2a<*d<BH@EbYR ^ U3٤>1R<'|TĨ.(fBucUu7.хY]ٍdXMj^]Uv`-U}Kr]rּ:Ųe$W2~2wָTש5e}sJ'auafFS/x4b`)296Q ?+q';_YiuS /%_$Sq?spNOyߡQ-ûOϣ@ތ:@kwӧ;J C^SoF6Qw`ͧpNT'ZSRgE b5i*٩[0x;*]Mغ'D9 M{S'(7y;+6bu5eaS&4,52 Ppvߒg(68ӛ(TQHϣ7Pjinj'TdgY2E{*Աݦ۞\vS&tw> '&W;:>Q33lg]Ԫks8>GcZOjȶ&BjZCṅcc\kZ0"~ jlB$f: e̸\a9pBgG`CY:,Ca[Q55PߪePcܹ20lK$(nJe`>z7i}oO=G'X ʒ1oK`S7%+m$3޷}$RoBERu*iGG\mہ%Yfdxyf q,7ƶ< ]F2e)V\{2mO'eqv5Vv̓ Nbu C}ftIX 9Q6BUgĥ҈\.[N\Q6poC8H>42SIUZuہ{HH-Q/ P]k5yze2hb^mβc-eS29ڊ `5;w7w$!6Q뒨X} >D &H*kN;Ĵ8\ ua:@Sc3 (@_9 6Njn 8EUy= Oc8NOl [k v==ZԃLu 4S[{\+`Kcg尵"\}JwFASCK\)ڀKe4 Bp.>bcK :n:ECץTbF50oL(jhKRu𠔩eB5;~2Ǻǫ3Y]&q9 %ۂ֪v%񮘭?mbݴ$X4Q*E;U+ {9so/ _4. s>aaipUV,]8uK| Jܵ8XOIzOcJ ]Gbg29t |jqt! I?] AiV]Gw*$z ! GW޵ &ӝ_d5T骸T!tEt )yB(̶X35iW>;#W/?ʋM$v v`b8x>^Esu[6~Z4H'W: 2d:oJ}{]︺쇺Cs!\]ե."Q.D(\ ZOg,kWI7e\[*R.نCDg,iz8.~[0Y4,.r H3=Cz ^URU*+ZR]0,侊[H9s7 9i9R!*f s޵!*͞]L]!PÍ 5HԝaC昚fB36{l9Y786H6[uZuefܨpgz~"Bbn΢`\\%Z27Z=|JUr K;bv&.[Z˨.3Ϥ $ La57L[n$c[kf|:sFLq}D!2+z1F=6¦?A'c'C W1c~} tesѾAD"f 7"48Y:д"R:qtF2ϴn4ۋFCL]z1ls2|虆Cv ̓r.PS|D3[f>>Tz#Ơ|t7B?o>]7ξ ܛIɏM&݋Fb~vMj(״b7fI|lbCncT(܌U~v1'-nߧYcjlKL.:/sOPw1[l!S?ugξLuK~!. tK-o_fػlK]{s6*q7|3k'qⵓ4ERYԒ7~H eU2蘽nm;Bg ŵgo8j^xH:Je3g~"@?&VƳB96 }.momZзVn} [9-n>ֱIMe&6#>Ydaz+tB5SRf7M'чa~LdMfvg[KAr({&4~vQW!0 6JUL]%cdעHTԁ+S#Om|Z͏P8{\cSNNv!_`I`ŽsB BqR|K(]sP8&X nK^>w+^>,bZ(3n g^P,t~̎xщ" ;,A'n !J$x"مԷC|OEt`{NV vhΣ6Iu(f eP)`:) HF)&& j" CZԏp"`eZ JL$:ԊBFf"€9ɁURYA$Grh+)Be8Ek~Gɨ$ò"N:)Ѡ_OTqM5i~7ঢ]?atJ5Ytc83v +'9ިPpHm8R35(q`\/2¸@YQ@)\]?@"/ O^-$rYTb/C ,m3=lk\8U3n2Q3VL7iy12FzXn>)X7`m&ʑ*=;Pq?%˙rXn!`+ul_IbGFpir .e4hހo.q/,HqxcXjqYa-&i2._%l6~#No!<|Q́)zc]]}ىISE=Fy/z̋T1=j~ Og޽%[5W>j8V@uZkFJNiѠ'kp'~yrc dBeJ2utl|ŘB\5>n9W R~CX ɧOG N4 $qS$2$pCD rX~_D,v .]]Ja0tzYÎC lM0r (lŊtӨ&]>Ņ'CMF~yB}ZESr5OΡRu2]MID `&j늼v&W#V)A-]}b&b>,e { RxZ:SKtF8P?9ίR8_ƞx_x ydX;t1 O#Xf__7輜] "?Al7QY Ptw-Ed9!R!D̓~>iyJBK`d4637f+2IpW,PAer,K2JW+D$7 ""cAjs ɧ?GfW3 BF(N_x=' J~J)M#%l  -K<,emh g$,^Â4}(o#Q޲1/UmhuQ{2XV `Mgqxͭ,WZi腂E\.@nf#ng!co}=Njmxn*#nӍQg1!lذi<9$ IT q6l2t抇 <Q!Qd1).MQ*̲8deŲ,l6AP4\[=6Fi/bؒؽB,#|mipȆV{m4PzvVNLľ G;p|r+$#o&tq]hìA]狇Ŵգlg쪭uzx G 6UfyBy_AkA\9ׇu\uY`&)@[0%q]A j96YW H [PJ(ۂP=i ezK6}vT^ev+۪_wV=K>owmzR˷ oO+|GJzz[mO4B T* TlOxbGB[oO4B 4X%4؂`{BHh*>pGB-V 4ڞiF;mA\%TnAܞP4B厄5.RBOj:)n&Dgusx &|R :R:A8L\D~hvYg$(|؇w3G]z$VuTwvL@#­g{MuDLvqcG](i<(8F͒$/?Tн?%3 *'F:|8#E=)}.%8 A9i?MaѪn|6+C;0jHCӲHQp sPF3?'T<)G:86h͠Ɛ Cح5@{բ.+4D-y)D3F~ jLVh/VDH0h>@+ԖTR ~$PECPM<;8WN@TկZЫƣ4=JyA ?PiV>Z> w'v=R*gҵ5+mBG q~z8WX-g<=a2CMa2wf%-m:u~z2,4~=pwδ76ēx|9,N ˧U^&c}`T+ϧ#xw1Fĝ c}sT>q3z#|SدM ~WdOIsE¿牊Hh0Ce|T s<8]B2 (5`Xؓ!$t= l'jA=m|y7F<T/zڝSa o>U~gfY="Sؽxh)| Tbqm齆 MĬYTd'z7& @S <8xs0=ZzVKȳ\ȍqljy5,6g 'SM#Mף)Sv2Uy`V8$ j9YmLCW(j=*5?A8b줋8(dps %!Qq#K5Sv:xHd*GW/[cJɚrᑾ@05p34L$$U2ph q0xd)kJE  sb4D@+@y j|é5k%S/:$Į`-\ j8 rDÚP{p<8/|8V@/ ][/G{P%>4&C #x kC;&o+PY]ҕó?C(1{>,8zܣOv-FzFF2?zܡx{YW 9{])Fw ^S;3`6JH_#<pec#WaKV%?ZJdR\D{NZ,Yc=2)*cQb4˒,2nV,PxhLg_k96WE|Ϊ |,GLF ɕ&dZP$Wp۰50obJFѵ_;*g5*7Eyv}炜o)E^Tq|Z!b-2鲨f _}++d}#[rwpМqTb7fs8ږY۩?iiG_x1~Ọp8gFvР'dBDz_젛L^SsRꨥWm}J@.X%(:,ݡQǟщ峣I0.M{"1*\9/-]Ya1Hs֫Hը k{W[-qc:_Ʋ uOgjÄ\*CoUZ0^GFQdl(E *GUe#!N{"?z>c"~JT?tM,X{ 9??3;jKn\Rٖ\|bӢgh9gM6] 4Pf7H5O1p> +yyM&Z`fƕ-XH& Y)zdPulTGLIx3\sLE/?3~U2c2M dwmPh /PyثLl!)3)3[0dNpO"o&ec|3y@%NSTPl67ϕP~.5:ʀ]u=u mr`ob2utp< P!L6~xw0=Ն)"oJJӤ.J;M5I7QIjǯӯ_Ǿ?cm|?ꮼՓrlW]>kr.տ~;.Jy]ꯓ(~swW'>r-^^{&^W#/Կ_@jzn7bW=[w'v/:v%2ݏgSd ; =mgO?. K?zvr.hcf?tM>Ad|7]prcqWz^rBp~M:E;=EY~x;v:T0*g¼ו(Wcp4;%vRcqrl} ˩hc}E80UJ֒lLYYشwY"+w3?ŽOҏYkC0ݑPDj&1Q1aYbKaD)7G]C͔Ƭs^:ZeILik:ӥ;i28.?achX uJ'`rt]I8#/E[T !%p", EFqJ">0dF΢šr ~lܲڳgUIKCf\hUGZaZHk{XmzOR "'sh/-D%y~֫D b(ΈkJ/K콄}1N5 n,7Y=@blHBx^kя0b4U+g g|`6zZX:[?Ilb(0tܿЮ+G1@'B'9*"i, 8ȲOA7!dtr1ÎMBZAȚΈIeI tv>͖$8y5T7j9:|~תKmb^vz,tVsJԬ4a۩8PV81-5 6_I3S)q .Up{NQ8e@p.ipJ?"3gGvbZxC`^# x |+?Xb]n'8?^qjjv xԘudU\tKgkn$ƞFJSxCдRIx/3Q'Y ~pޟ2QVRLKX*ϫ.K1$ȩW<9o~ OL`j XQ1-͢r ϫNxfY;Y\sz8H\q~&(ܕKכkv%Appɇ5p}af-h49",ɭ 9_$4p7xL֚EbҜÚ6fpkcMϼ_GXRMc'aI4&9`T cXD, r T_RUk2;B%dړܢS=U]fЃL+Wi1BҘsM1B"KӎXo7wr0cwwuՊv{& 8#{ĪĘ!]+E[˒3ÃMH 2etv2~׏$E^ge˜+浜.T_hš^[YksB[%K@% ƒcâ%@Kb5KwE?~%qpp(]!;3KJAQe44-x? ͥKyLK?WK>1k2if}PKfDJ&VزA1hݲsAklal6Uy޻FjT Ȭw5d o\|58 >JJĤJ1=ܻ E)3d9Q* c./C pvNUgMxΊљd &)#Kc[Rkt!,mz%Q|EYdG)\D(ʁV}21XӇ,'h5ʽVa x 3X:h tvQ5 3.!K)H_s%ջR,*?+dϒ2r5jSr|%G<BX:2-e8 ^oYWk+Ͳzɨc7o+";.Ы5BueRct~a3Ujll MXK){t  ܫ}iى/gTIp\ w)j kP VϻTBGƮ1 >q˪7֦kCq{^.zP# wT%=p\mޓ_u‰W%{pӱUwVlmcRvD#P3sAnXQQB +cl'{Y/=:!eQU~,pTy,4 (Ś31%"Oa1f~'. UL" , r A!%-zKgW$,у{s:IkB NmfĬ1Y*D$-DX: ,qL0]aV5Sb-EGwp3U/Df7sΘNܐ"9'a81Sݲ<wр dr4$fpåw̳2/3i|8\B("rߧ&& g^/<5 Ӱ5n38{, U$8Ďt8!ppN#AyY$ #;n羸kF?{?~0aէf?dQ |;D`>ȫ).G-A\!tx >zgg_^&x'vWjaQx%ΡբTJ[>xaǭ+z?e?5;\%HD@7Yh̛oxo5% WlW$jp.j!>|^{ySdQ׃8.<*&:k|N#QEM/"G^k=Fp$Hԡ'X#|ʹ5 y/Tzv.(!B2aM}_׃'1}sKJb,^d'0m03#+hc[-_Bk/);XI>'%?:g(f{ː`k$ێv[?՞o֫LHYTZkƮ1-:GK,QG0m&߳&UTpyumj]6*kr5i~ yکkһܸ{sy 8/l<>cSaK*D2%X֤)k\ȎE%vwcdd1pj% e!bZ_:7;mNVsAg!8T2G{e}z~i 3bҦݲF5ˇE!7;cZF>[ċV`($=%o-KIöPC&Q=G)aTFL8>h& bpumF*+A˒Y`i0f1zň H/\v})ЁVP;%#řsLzn5Qb7,)t.DJrS#nr'8b!I(]MƅpY $Wâ3-;4=Yr..^)2c X j+mKdvaJ5,v14W/oֽb2)JWz0\l`z/m{>QMx#b, '$JXTD\2&ͥ :,(|i!s$f ^0F1+5ioTheTi2aB3Y@BsKg)ihIQ݄!515zGQE83q'Q^H5q R-*;w+x/N]kvI=s9 )Hmc?H@̯j0BFwyA `<ݷ%h.[&Yհ9X#1H3O ݯPtC|/p$ "=cqpj nZDd0Į/CxHpzQ.x1^lx[0ʽZqƌUy`֗tbƶj(t.(o!qpإJLkH`86p/fI.L(wV[oI&﬋86];TPҽْ9JRQșػFndW6|?A0nv`d2;&v#UK3qoeQ=z౥nxX:,U΅\p8߯3;֓y d+8rܪzx5)Uw\pSRmG%yד$Px~g4l8nѵ\$zIS=Y6{N[ig~; GH> -;S` <(ٛZNA$yq]緯&LG%QTmAOZ"-ߘꥄz?]b0/sOIdW˰24b\Z~4WI?ѾTZI9pIhd(FR1ea1A1p0_ iTv'k, G7]#Y-!1fbC|/ޝ5<Իe4DAYx B8@Tr{w8Zg[? Bgj;.0FYtK[syoJ~G/>`z~)/*K ?ќ,כ c8 wyw5HnMI$IDe5fXK'|yp%]'EFD;kpJ(*W0\}5j>_Gj:{~{j. ^OS ο3y R7~w(6&w~&@鍮 ҵLĬǵuLns OzyzR+IUFq9HxW-\ "_Z8_C.++7.Kz;Z}W{3_N5;چc޺JT(O\77Em"f(mm$dfWi f FdY'Hȗbqγ}v~|0I/UhlFQ࿉^'VT}3 WVdχƐjO PRc_t5”|h"t35 nDZ(yw>e…N |n?Ո qQ̽@q8\PJ68K *d $ase&3 #k-u"F'c>2ziYý;wmoL|VnGpkZ/˹^cQZ I!Wjb+,. }Զ'^GhSZX'ĥ5AZ\ڰ]D>x| /ULNʕy1~x7_]^q)\5=ah% fy:ƺStFT_E-W(b[=mˡF<wrW1~:ϻW7jtf@}]${bOHx4fU#$ڿ:@_$#dOAٴ{AE'^xz p&/7zSuPc&΄EK).(9 -`ܾ/ՄHw蟅Pс{$t.jե/ f9+CɃB ,"DAxy2FwAcq[w@PD"NJݹI1`ڦ O7W?_n}%%* W?(/u,^\ f3k K:SL4_34Bsr*0!>HZTu凒u%IM+*o G 6P:{p+-]GWd͐Fьe: :#D"єqSZ҅\F-21 0]Y3VH3Vtbj֌eX3:/&\d4 x<)NeNQW)8]l̋JM+jM+oB2~86ca:>@PLP f)VcãrEhG -}ޡVc:szsփZh<`Cji{A'z'@ɻҊ 6}AI}+aNo`BozM>:$^+ \tm. T҃Zkw$+4%7x:BjNy7)%a"퍯rZ)Ғ0Q4/׾X1 wvՕ8)P JK=^)15aa1`F2]` QU ryEBG@vM/o$ Mq޸p~ fD Tӷn-\X٩ĢKOmO?~׭}i[:DKb;o:a&OU|s7F ZpZ*3,M̰ԑWM˨a $Y6Jj7ڑY4 WAi/^t4Bty%l8l(Aja@W\]rYtj s I6 WpIwJׁoZ`3nZ~֮ˮyWz{l{0ywl}~3S{(0ќyqPhs]-g5](n?o{(2aU ஄mX')|O3I{_YK`NWgά'd[\YðI ~ޫ3zŎeZx!ů|@Vd)(B>^2vahtƄ,iRK -ˤ<`_?Ь+NLRW!u& Ɇpma^oK MPH(.ad`DC**.Y]ZIe+Te"e Q[2YUʳfƻ5[\K`gK~gHy(Ynm.i5D>:|5ޝ._ 8cXWcFydҥFyÜJ5%FJeވNǎ|P ZUU|5#I,p{,FX>1f&oinì^,b6sbƐTIh=3X:ɸ'A0΁MmAeAf IcLX &F4|i㬆+ƒB֝jl<0 WSc2% F&DkL<vϷ[s_SmzL;BfGa:ٰ;K3:esE dXIyx&NwήI9!pj>WFSD9QFԸ4M1W"%TL:`Ɇ@gt(q@%7̤4z *4ԖQZʅaHZbQiBk ,W@Bh|pi+7&SUV*-)1qƵM}qbʹ0mjuv<'n:О4Sq.R 6ԡ sDpa%9Ei">r8CxƱ.s[YՓ9c٭3C F ӓo:XX 8BL4;J i2e+n Z:| TOKŨa _o[EWuJcCQ£f&\G#%zpHIqE۫xEYb! o,&NX @*ܜ4%$h|-RSqV?h-z?eRV#%ba8JU3ˈP"V )ZLZTĕ}Ђ )[ߺvF1Su -8g?dK EFtF%,0;-P5Q3eE 0`%Jx/Ֆeͤ-S?nX\Ѳݩrk 9ݩy6U(}zgtH;fҺ&Z $ Gg[BzF-brij3EHk>0)7`z[h5zu"gPUQc_/;RnhzW"c:AGG"Gv-L67hqDelQp{MNzc2nQ#uk1'>09 AJS5-g@r4Î9gݙ03ֈ0#i[ޖLkx~pjyt9AeUԞjxa@DY)m{ )[J/_p pUk yU.ܯ)J>4Z֣ h<ΰ%\~IGu}|vLosXDġo \ħ߽,6ޗ'C&ΥkPTbIƓE>2r+Zh0o)/.-YF//43IƩt$Dj#J"7` .f{{nTƾm߼rЍӛ |&߯fW*^nR"փo^/_J3C]pPTJF1UZg%< JuJ!:_fM5p̐$p`pZ9zw&ϯ.fٌ9?~F =Ln`zԯ7Ӽb4,bL, znثhڇ}P|5>`^4T=Mz|qu[&eP\N%Lz~mZ<L iwnˋ^h k'=w=u+U;:52b|-?߿]bn;m7[Z~ `K 84`p' 7*Pc?+>}B '8Qj?_ofQ-`sguڡLDyn.?B.o9?qm)M~KwE(|lCub`vςf "b}A O<y&iѯߖU[Y?Pփo2JR{~QNR$e>/a(t9x"7yK{?ت/GQtrrBtŖ/`T#=ɷ3 JXYJE ]5*S my!b71g+1"C;|ކJ#ւ$Wm< _YK0VBl'NJX6Єg +Fn\M'($NT,j+z@l:Oʙ +hv ]ɪ`~b_6CRܥJP€)+U(Q*5,dC&%FHX7StF*9?<GCC(.$," !U5$ Kِ,Q(YKdDzMyS%te!g[{wc ~^ }6U5:XZ9:/5E?U /aptyu3?ܺj+ 1zܥCcaФBF$SIT<8_,}}xaj.w^JM8* ųoJt|\B$NR :sFk=߿zu;_&bM'񡲪]e2ɴ?!h}榡󯽳}!&δMB)/gבMFn6RXnU!mx~V;?$ |D-kPBD(,^(Z| gwWpφB tIG4 }.h^ƕ3DENc%ݽIheyF(g[y}2ՃVůX0bZk$v/؋jQt4D 'Sl&oE~c5 ,U®YK7g%Ń{7B.>^U}P\Q z?/k)!rS*$-xIy9/Kc`4}ws]W\y|*jB(F㼸)jWYƠ1~)6׷w+-%` qZ]-$AnFEw.Yj>|UEdx]o0:?uEX[$ oE.O~1Qč4!"N%ьN %"&i"̨ʘԱ(6>^n`!xVi1<Ԃ2idqr]eJ $)|vϡobo~OݏR_% nYR? dqW z8lz¾ʼ'ت] hY}H~n?/d\][Q,kp=sBx z9Ye*?Ge̦e$i>u?WoW.t R^騏_F/j NƌQO]M4_\|5| hi# +O t`]e;j7*Bo1o,5t1.9FMuWI 2g֜I-$gJF,PJ҄,4JiuGrԹ7]6e?Q?QBz}C=RoEcH\[ԤV#_W'Sq,Alo8e;+cBӹby}1#BaLJHaӜ'_Ŷ婻1j4D$lۂ925eMP`ńEFtjdE<6S_cj+Tj\CNFkL>QosT $.rJhiKLHjkm$Ip^lm1ȧbQdIQ,U1gЫȈT&w,Lt)L"6eY#Jqcĝθ?AA&ưk OM3גk!N%vvZc26o-cy26o?ܡOO @_4ճˮf[zћۿ5-'~y/\F|e17[uy4Q[͏]7Kf8Op-?=loZ!ӿxOvV=G͜s?5႙zPQ^FK,THZAGY $ڤKVPq!hQ*45H,ZJ9,0g`hCEFB_O?sѻ4o߭6 zo=DpcG>[-v׿?6.=aA' 2jp;-ge{1;8|owqмG&j$' 'k} wGr 8 pv~(d%񥵤$USa'IYDv9x:1ǂj+˗qy$PXEk|s"8N:Js!d;OwH`9phUXKET$h)&X_Y !4qcєUc RR*%R'J|03\z1Юɀ"ӡ[*5-ð[0W]2,H V,'mp0ƽOط1;Qcvj× E:R֜4Jyd 6s+IC߁G>jA|ϛCf#bMЩݺa4V{)3s:i$?P0ł.M3PFEwBC$t"ߟ6\GK8cg m['ZYmDTcv2rcg|}[!qR4Dz0˚D0`% =2cxt Ǒs9(GFžJM%I0r [xĂ70l"j%@^nCP|,b<@^x\M&M"5yEBڅlV}u,!m#?/dw(+bVd gLe9`R9ФGF݁@Lg! NHcE*AS,s 08A6AemT-ޖGê>|$Nta%yxۦeM0.C1tqŌ#P&l>pQRI8T >~^[j3u/G!' yw+ p@#5ɱmXtQ|&/p\֤,Ǝ6xbDxԀ֗5QcAvz[')$K 1]E>R5_0uusLGF́ mg3,/R8*C&g<^Q0?f H gcJcUX$tEB܏&rJLp.em)I1Nzd;Qc!l6ȓ\>}`n`zdT$,3 gYaeU,i o`vJ\X^g20ТBkx5%j 5HΧx5|!p.RsEXy|`F.HNG*~ǑU! όx+"D=2j +G$OkE]vRC#v)# Q2@+V.!< ڣA V,֛A٥({V8S5Ok5i&XEm@Q!ޭGFŵnQSMr HӔSm![OK;*Ɛuݽ<0Q) ht+_,rdy4u )^H"`H !Z`A iSL)E(+] {#7K:`0,,l.R5flE>_;_pYmcPFmK ˷̀ Q,>Δbe}iU\GTװRORk#Ƃa _[By͓9h!GB" 0 'O6_Glf^qrbk?җQٹ=2*Еr6r $"0RS@\e\qeT8Q tߘ6ϕzƾ`a SG4D& 8B%R=侥~v}vn+Bvcg\^ϧ˯tR4jE3QMg//݄zQV(bFqw]T[p|@ i#(j>*v5zyRZ[Iw0i1Xn?*ul.ft߶~A/7f~u2nO+pW0}z|%*~lj]!WgۿTת Eܧط4P'w5^AG>^cif4D/ E-[wq E3[v2ԐQ7 CC\aMs9chpWGۿ~{ 6 ~?54;UaxBwݔb ?@o)qד3Xf.Q i#&?%WXU楝t*yM9v WV +_aw6qjzLMzuD4ݑ+(L9S/핌lgW:װe | s;݃ a׹ 3)S{[qG{!`qb'\N1T8*-dU@O*jffƫK/N(gNO?*Zb.U=XYȭ=b zSsE܁b { y R =cgkcY<>rsHPe#{g\;|Ү9OL.豏pM|WD<QT2y7/2zž_W۬qn'wlj雒*otp<~28udRοߊE>[fnb q3]1Irp oIp4XѶ E Eܲ xq #'΍tUy~XcMq5;Q7i5v®^؆m ;k{Z/t!w|B3|[.iN( vmvI L0o Ψ/4]E>rC' E67!!t \n!pо}E>o7]A֔+pYߤIAB̊xp"~1;Q#dj:<+iBE*wQ<˩mW.ove,Qy]gcIuNGqϒB]}rE>]~ v;"bp"N23MCQ7nCo b\XhMDE {7(_sb&ʯ4GFVPVJ4mfחZb5xbX9وG! XttH1Sћh6|f$HF"RA:|[+1kEB+\Uo¸Ԑ)N> Y%8}3DFwB$ j=$L QF0"ss!iɕ8d( C>rcy C>Bb,A7{y{YTlҩe=QI:6 \9TwRZp$Q ;eNi7"QMJr( x >Μltq1|,vÀ=*}Gr 8򱐛ԏ0enLց:Pc!燇/CITScm12BDw/Wn &{ɋVӯEsI yKXj?j2J'm܉rã9|,ߨZ ض_'15{GȨE]'LU<~`Дȑ!6k8Uqr+b4tZ}e0 JQp>_⩎1 1&1~~ ΐ^EXԲ}b,$WQlIR{)" Մɚ9Z6XhSM 3Gj' _ ?H#1f+؄NX`xvg0úu(#ʳP)Nof՘_}WMo{)<~ߎ b\s15=n FNiXȅƞ@az;T%j6Ez>Q%FLYz!EO_lۉk̠n,|$̵ 8,5yXs%hRQca|x/n&_ōfan:L b\7Ȇ"c5W35n9of/iylO?nf{}LcȨ!|Gĝp* V] <wj%55/:@c/ $~cZ:|LvX|g8!6L0x`j m>C)c< |qI]9I] Y0su[^V]]][.G/3^V0rɾ&} o4}Q6 wx_ LXpyY·/v~ص3mb'ڄL1K6unm@ -F&_pW xw&s3cj,K>B?wOخeU9!,%C[~{CMXrhT*n3*n9ByX͂+οye R6B a'2[Aۂ3T#(`, \#!Cob*<] YePpњ1%-I?6ACc g!{Y/y5/ wVnR_ZsA,7Ex%^bn2-`sHsdKs*˳U;<`?yIw&/}IEluwL(n ئ,f̅\F67>aU Dž4S"d[Euˋo7ܽc"'(7r k$p^\& wE=l_.9< LXsFx̔P;?L\Ӂ [TCP52~L.>~W_.q}aS(4A0F8#`~#˃;F]aʃ -U1eU~͠G8ɰ~ ƾ ӇΫB! C~5+Կlu\;0?_e$P~zf՟'U5ѿzo oas]""`iEဂyD.ž ?k:8t#LN)aqJ-WZ"&ilnkӚ9]K(xJd̡s]GA.| :ZAk% hE:m̴dVl E}JT}s9Q(MSZH[!Ta{$uҠ˿ ڹɘe{wv&1קZ<$A -C~meIu ROIC}JZ7~sibs7WiuXc0t xJnGgQ5&|wG kiB};kQBCY@&zGAwA4zCfK8^^A7z %e{̜"_wـC%Z1DC_z#s4w_^b3wbsCP2(ef8 #mkV $_+H0T+2f( |GNы,QPnC`V(ν|vߩΆL G+A jt_EU)Af>qQ,4NJMӻv(pO@8J{ASeD PIE¾XXdXCn\r|K 5j*2f1ja4EMϼ$9! dA( T`6{ )1 b%Wmhfz.Gad!)z!sqSSQoRЧG56irf sSu&j=;ottG~q8vdjd^|<Þ?CZZҏ8J(yDdft-F~NơNׂ`l |_c >vdQ&\xiC%afaE>QvÉTO^-лn|WL'}=<ݙs@0&9X9K"N!3vj^`)o~{|ХH~]x%ɬwM2tﻉV6DJGTx _g[߭vUZuz C deG* d!FߴTCc)hϺTKeZʚ3 j ta^J3_>vU:yWmy|xE09Z1s p_]Y^m5FO*#GdLވ p$WP|rx`v/$2Oމ V9s0Ν,QOH%CTET۰ m,UEY䧃+ڝ2c>S#r\Xaqo2; 3Gy8sXFDc=)=9zU%,LaV"sޯlD. hE,a^)_ٿby] 1`NFp su:<6`M$?2x+?PVO6vz2X)dtFQNCC 0||B=ggq2f9VY*8JǑckND41:ؿP귣9?>-w'Z$qb+i {2l.bPB"lP U|IZA7|{%+LE;LdcJ$ Ä!.BlusXKЍCӒyS1sM󚺪5]Hd̢%VGK gKi=¹D~}n&Hvȏ?6nV?A%a:̌؜Z#fG8B"E2C(ꚺy;iAYΉ++)Sd "c'D dzF$H̺:27e)a>Xô*L0X|I,׻#NzKU>Z4=bK.dh0=~7Vzfʘ݊}seNP3iTȠSa2w.5SR5ӥ/"6H2]O 0[w±˶OX{DPMhXr,U{:}JBj>#QF&'`v [y+BysQxܤc$jL;,P4a9da7<쥨(HP ߞA71z4T7b|fHs$ ?(V<@W a!)E8mdIq5t@d3Q,KPzcXsEj܂ V9]΍[BkYQFd钶hNfrьKD>>d?NE@|R܅͛d?s.栦91N, /sCem5dlEݕĝuA8PϢ=1%yIM{:/`/xҿvXު` Hļ *DJ46KۧT*ٷNn< | 0Tdaa$6˳%fHh́F0is@.Jm5J/0Y_bflj~8\Ru3csTQ/Hl\Rsd"2~+w.-+.x9 i4oe3~t"\licZY_/ّ8 rQ_@`G{ ~Hbc2F74E"Q"R?fJ |A4}v} '8KH{6v|D 1;!(AD٪;g)xy*%8lzʳyf6!s>pxbGc3@ 0ml!/wԹ0B@4 m;dRc/N6[sZfDYn%ovl~]@Du:42)U*Q'&i=sGǑ kL\oʵ kfo c3 F6s~?'au71?|36"/ 0V_Ro|vI/'ښƭ_a%I&q!d0VL@[-i$vTPז֡D@.d2٨GG>j-v»+>YhVr{v`\uvCڒ =*KXTP8ϴT^)|ӧ: {Y23ZW5{e<lik޹B%߻ܳ~x^P3:YP^3^zbש vvBtvx-b{}]=ŝ 3~8Ku|u>.Xl$^ڴկFJɕsܚЯՄUsE:hpp]C  ݯ%!&Ǻ SK"ť_ |Q[FJF-hGG^{ȐWUg˯y 79jo)Fbʦv#C$oKLA?Or~1?}Pj-5ڋsܿg w:\<iOJ-I+(9Y-%TLEkj0tdنT0½J.iFi1z]n<55X}K<F+Pnӱ󒔶$RVyɏs46ϛGaI,|ЄO~Y"vu:iPq6J`IWad!{5:y˸q2EoOA4рU8&>3G2yg#IBJXN@1io/ mWw;fhl߆ٴ}m}H;]G?_'`W#q$ Ky#Fdpʽ F qAMpVȞ[[|˺EOGi? JƁ Sqqr!pYR..-9l *(T- X#%{FFW#Ԍ'KM.BP=%SHg6++޼A3c ;I*2ZzN_Wَgn-.v;$-AonI3q^Oi;U%/Q1o҃\>`TqG ˔(9&&a j |+bB ؜ h)>z ɯW[v]x]g 0׋KK堠8.JK3XX 9؂pjB8;uayrI?z9[PPF8|A]u,9Fdpy/ce`V Fp :TE#!YYsӺ6|;uЈ *#``1؅*2>DE =O72ui\)L |>J4"%kɊl '/ؖၼB5Dɰ PI֩ (Lp̆5)rk1ÃSA{$lblq9e U9T#?g~( {4{d@L.fP[#V `\+ rEEuvЈ 7ɟR(Bx^%,kuyAb 7rrGdKlǓ6O1GfVt%rfxn$eYiUrcA#"!//RZAimùi< 2ֳ.#(gN.{ZM|xoLJG@jtЈ ot'2ɸw c?HE D{nĢG,ΥdBthpTILXI:0iDgKk+^|VPb;˝{'A'B ؃Vq!+AprD%ۇ^1GGD[V4B'hIlTfJcu@H74^Du#67D•AphI!d+\uQ)ybOl*KS3ӏ|PqGER!4%/` ޤQ& wz U]VMo &YF\pLApWw/=Py4"S3b1[IZI]s@JmQԏP0tЈ 4izd|t(@!ߒ QPMAj&N&x#U4 P_z<,uЈ }xM\w@Td@O~PSPӁ3RV^+\4HsZ PO1v}BXP&ܑ`×B__z1vFܦ_BsE4dx,S"4]i4 Z޼A iplzPDeR{*aܱ4S7Y([ m{PSeAJ^䡔T5 -[<z${"1z]>]yCtvQ$A8xt1D]>wsF~5_֯K\T& R#Bk'ܥ9yJbwJYӲ&A#28lgy":hDG1/z*\TTd$P0ֺGMP#E P$hnt`꣝EF-B&Te*tЈ & U)rD\h𵅈**BU$7%0$`qxFT(.OeJ힦ō2))Y]W PoCT"KC8X&_w~2S,p4}=M8⏳i|kxj٬Y6U נ̭h}ԣg.a.tzUwt7c_j Ȟ'$N'ǭ9 jp ?0Mh%e.( m`+J[[*Y/_+:9xrj[6w*DDDIW&59'P;Y1|L -3QٱHwvr`ؚ [wq-.UŢ3o Qf .PxP- W^ӣ׬۲5G QPݮ2{y)Lff-PXO<#n)3V ~=M7MvN!aJ~VA[goi‡lV햍:g=l&[ʼnw|~zԇWdX 8_n~ѿW{:]fvCtKCj*hYqZ*e԰[ZӳQV[] /3X㸁Oญ'ͩ$yA!In,r/?dG;-ဟ~V8o0LBBʇǽ#w4_/a>W(ƛ=J; ѳanw츙%Sdc./#%y(q-:F-؂9UwF tCnXznlnGѲM.O |> !E[PL&PmUi0aG1띌{Q;H4ã\U5:ZN^-#0PtWLQ]5Q3} -ޏ bdIX-gx?SHBu[X)/IiKJ;/!T6|2{ 7!I]y'[6-6~q`I ߠJF2Paxo_rA_b/ 3knC=AY*G;E1ȢsE!Y&[tj!Q ;Dk&bmU W˸ZH5OdBEq~6{x|5h<[w1Gp0*%sNrbc hڑɭ?onJČ3OkA%d塳[š19LQQ-V7rAo'0rn[90 DrN f(ӎ`!!Ї\IQRH)f-PNÀIPDs9÷|;ɇMw:_ȘtN]g^Nvz8`'9@K鼦_#, VEL*;K5eNt7_zX)M.f3m˸ Gq-Cp\r+ʄ'"83.3e7t'ZzZk]d^4il[\onl靈e3ۗW;ߐOЪ~TomSohyS6 4Nӭ4F L .' f *`k U cJ#68R$#Z{1&1V|z8ͺ17KL𕖌^B::,5S fU1_iIlZ!kAʧ;ye M:{/߫|ozNoc'漜M/>'4"m]`nzkuH9EU'UKUJ(0B0">KZm>nEmfܨ6]?j),N92*!/>/ٚsɳ14"Ue-VӔcB{+Yc#9/7N Osٸٰr/>N!S ZsH`y[qŠAA!+ZɩI:$NH`EJg59mrIqvNVjD(SjkJ0r)ԏo~,60̦`NL<4 u?.4`+qCOiC00^jaG` EԦV7;Յی6ns^xVw̪A뙙nmiNxq6KN?"'̩u1E:գ\$2' ) y B񎷥t-[A7n yIf7$Yv]x+d*k]!%%Ir]V7KQ n>$8lM-t9qQıf9<7fZfz2s8\#QFW߻:'Vt=*`UEG9Mt>bpGZ(.ruAӝǵtŒNJ=P5ƦwV=VJo4)FKqF9/7܇4g F)/Z E F!.  Zf^irY*'h &dȻRX!4<0A)j&.WșAT4.^W62'Ԯ"r6ȹM b8X+\$JH4s}p]F{ӹwҫTճt 'o.}R̀>t,]]x^j//4yj k<)GXc=#(lr;rv5c̱b>ŗN>YڧZ9E$"_hYT4a˷r^4щck-1Rs{k>Vcu?r5r]o1Oz Mɲ E22Hjȍ1<Nh|pW/`-S{Q2'fJQg`kP!gh՞2#,R@bN+RbjUr }}YHΆaϋϟJ*%٪i82v1Du 7K!R^0iTa6Z a4;Mi6L"z#{XL \XQCW/׏i7׾@N47K'1VJ?bi_,mDCfZwӡ @s`-5i:~Iֱ,g!_Zs;DB'TJ ڇYC?INf&έd T&N0#rMN ֞>+>}^uMRh%?pT!N`{+%sʘôogHԚ.f1oFXy:M`[UKBӆƲt dd\yIQ5y2SvC'FVn^C MX0C/;q8y &9]tw$e|&9>9_đ=YY<[w"(999VTp}O_ZYT_17$BKE?Y.LWWyZ `t \| _3yo[Z<`PV112[,Ztl!6PD C.Ld~#18͇x"IU4?NƧgH%,Shg0 .VBF9 "2 Lɫ L7{ѼD8LQ)$Kvj*8oIO}8gv^C ҁ`h,^D[Koc'l0<.uT@ F gaTS<4]~t5]`,ὛFp<]LWHMq:/`az iO3djv2J[?h2!,sw>}~2C>}}{̯YGQAڳwJOE=اɛj}w~;b-X8w2ʒn<]!͹#{'՝s?؏ǘ'eX:;"Y)rU43q8zz&NGsW@Y۱%y]L:JqL/s2MDCt\R0+ F.lBFyOTlT`F, @+RhOſo!3by 1h`ۆ`KG,8 sh7J>;2LdDUoWFa^ xuc2O|Sߧ[e+0@8t: f7^OջCnpT;C0_&{;q5/ŝf 39363t~<,O¦a6݅)VgMn3jyW~;.Vvorjisr"L:scՕ?_z$^rm=7%igFk:Jl5hEmLjO0u}1sՄq4iָV׽,G?` 'nE+`{\4IRO.k)R!x܃{a\mǏLƲgy\^a:~1Ecw%Q߸R4jY%b?_DUv8`i|s~v1Kp3"ا}4^JJd 1_œ(a5yڦ\1a=Sw2AdU܍Khފ|}:,/E{3ٛVcV/b+LgUq\[7b`[ٖIA#'Oo9X_a7l!k~zU1Z9X[cR;ĬW逰 ߜ <Xe0bziHSdvufՎz%l2nS6#@A5vՎ+k,QH AOPy029ctb[N_Ft0«NMݞh0bQIY+냉KM ` Xy$J9ő> 6> I]Lw|Yu!lYvaNkzJ0m>ڧwn4 7E 62(0*GT KIwku3g gϖ7Yn*~ ux-|sw td`~%dm&S/-{ fZLݵ`DV 6(˃ϥ){y ;UHs sUTTTjen:RLb K-{H /:ZhU4%;Erfd47ej8aj` 1X-(CgeNmK3Z~5k`kY】\L@1%`8) {aRE%u(`\+ܖ.Wc['CdkㅑJ b1B;,YR_0ۂ'=pQ=5]dZQ5Ps?p]xxuknrF2tɧK]~:w0YCu+}…^ɣh}dAam lRyid"DZة aQKMLª1jveASA!a8z9Kӑ #k#XnPRr᮰Tls'4qgH M¡X&:XqH :q` OTGQ2oxX5ʣdbu2d,X s&2+`#1D…4]q؄6@ i#g{B2RޅSE~E ?Y>mt5伻t{xN{ )Ku_Mp;Lkr0_$7]:]~YYp8X/Qq!̧`)/wJ6ԻR7ڼgL>-;GqNwjiY/QVhy4*,'t7,|(*,J,qEKE6^JW2]fMJ/Ӓ#8 }T贛Oqghh˚ .X{y6"%O(RQ'*pksYO(P Yz ],%,mEiFg#9LxGގS(ɸsYb+k)sLtlVG[z>zvHzXEφ=i)LZ>z(LN˫t2oWPpkkrEdAa3J&l!qV,K={k{RS%[$Ք1U yl:cRSq$Ȭ-̀$W4OE6U41Px(pfT~l]A}(7 ̊S'cXD+6I\.@ydB woCy$C#E--ydԮvi$ُ cSN).RC+'ƀ=@`X)f^aYΊpY6a1Wͅ@rb1@ރ/WrU>{z# '.{:<@ B@@i@XRamUݷ_qeAPb1&fˑaaP= 9e.G@ eX)*HVA5jF@h&iKL*E<UUf1 %#\@r)KQΡלA]_O3EVQ> \Q_[W8]q/Gd>0Jz $P 0k"s"ۮ(˔ f!A9*|)LɂB A0̙:@Azg "m3"myH[z4em@;}.{n8kϥzy^I8Ssc F Bװ[R} X0.2dYxkWRh " wzIL!sJXPIÆbϱ# Pj,ojzJlx@_[?_wiҟ6TR X6ﮞPwB,i^!ĕH{BY- #2xYNt\B݅1#U3{6Z޳Ms쿮יTCIqDDũy>G.;6 9`xd-"fQ+YzFEIvUDbҫ$gE 95QfRٕ\p}MY@[m(3۹tvYT eON5[EU /|  X[qڄNq|m~ըm+cٞ%=XocQ Ùz|M3ސAA,e%Ѱ[521eoO xm! E7R|v71!B*{RNWu F} Zx 5cط&W?'h0㐍ӳfzӳs劢Cϵce-3/.W(F^fLfSN).*'ƀ=@`X)f^aYz^7a1W@rל1끜R!&+FVjQ/#hEQ{x TgLmLZ=TxCWԻ4,e_ 0`Cmx݆u(DռT/„0mIiU"tdC RRJ1"` 8^^6rp2~v>I)ٱazհ:LG~VqDpXj~k/l>9P,e ܻhƂi<P 0(0RlTc}n4=)-ai&8;( yb5b4'3Tܽ`rhw5NI{rT֯wɢx9{Z,%۪V'|FbHLEz:K~\LVz|LB7#.DOd15l 1=rs_{- }Y)Y&OPaHNBQ's`}'ЦbWwYTv&BS9 0'x|>@T6a6JRW+9 FCC @ B(1WQXrA_h~*lo&.Vh<]ZOgԥ l 4qr?|j@LSwq/i Uz@%F\[>'*ς;L1Bz0BB0b?uq)G{)dILky_.ϔ1$e( O_7oôj2A+/.1 !Vّ\2QsJ~I8-"K*Po%z?"h۸u я_=|퍖zDw?IX.>M&fytf t?hi=xZ}lEs|BW_Wv==/̦=gGX 8L1~׬~6I.g1~SPaXw#.b a/ǍvQE%eRɶ{2\( hX^oy%7d@L޽ScG 8KANh xg5eeմe2%Wdn\51Pvnrt(XoR\W^'sJNVVR6D GS@@XѨ= 5hvWe^ڴYB'ȋ3)]Vsv1OB| : I`kyjjN19"ryoyG_xFӌ*216W9XFۏ >C{<?6ؓĖޔɏd2[NbKuuɓs}S%&ɧ?;)(-n8DӣZa\*'`V@D4RRpwem,~I;eվCbc`rSp*3I%>lKlIݔ*2f,U9ue=h'RLh*"6FB)+gNHZQPp I;֕^WMy=TC4Qǽ Z)aam>)K–DUĂ -M ߒP%BSjKJ;Otu^'] XSl(8@-W`ϩ/:t,:6q'$`#qH"d`,lBX,TdFUܲcHr8\]i}2r^6Bz㎡; *޺fo`=@u1۸D?uc I1` !A6O9W%:wi'wv8Rp%<"vɪ:Vl{-@1 U@(Ldze br~C,$ r$2rJFi(,JlU-z{&*=hߥB OD-Paj>C S$L-Ln׌8g_hvq5]vQS)+" 5<-r ETV]^ϙIqr'Wk'צx\BPg!\Jls־G~F/L{\ڳP@f&{?2Su[|;j$HR"| cN؂Պ\wT<;|s:L{NĈv!Ă2S`rf耥ڄMg ~)W=jlD][g+`䶻nU[52B1ȸo5ɁQ:EKTx,(Av!k=cA+o=ӺRCb35L{[LϾػ6uS~[ͻ ؼFݪpsg1@Ҿ}7gPPԎ{c[-&T}>lٝ@|hzʐ0)́gXCM~t5CNE=n*U1LjxK8JToMp+n;^4ې}CӕI /ow7Xb(iW];ԃ}cl6~}]փчCI2i^H qLN`bq&N s4;tctkjx<حZ~ic}.I۵>xnqˢcC=dy78bv}CZ1V?4^Kf~0<4Lwnȥy 1)=7p*0|ۭz,vٵ=|>!xb0qF'N s/~7wv9pɋ_vaF.t.vaPd^}=ΡDa dkMr+Бg4G2؃%sM ["6ӧ#^~J}G@g4I1Tŝ6LX׉( UIcsީN(&{ڨK9+ - -11=,L@yˑVwߋ?W+\׹N㩟}=.;ygpmhKGH%ۣ:1iΣup}Q/,({۵ړGR6v.4yb*Fj&ㄹsp M8a)5rTvZ$_ ,M~j [(2Bi</?0ce(6 򚌏"iɟBV 5'҉w!GA/`ϻp#IJg7𫛜{lwgFʭk2=_w׫̙BrkaAuHPЭBd 6*I$nha'v-CѮڵy Sox:^r7a ӷI\Mf>?(GLGK/ bryε:oyt'Y63z7`6fo&XUʞFQGf6ZΖ5 lB<1uRӘ\͙`=3iV'cg2 O0T|Sro$Z[쐳"'=v[:l}!1L` 0!eLXό'!JʘĉL<bO1Zks D5Lk`_nI;⛇7㨹\gY++yS6i{7Mb]KkDp'υA᳼yQE9 JT3ET]4;*XhNi`Fɕ0:z&|BI ,pn ̙CNX`"3F+BkkG% )j#f9'ȕ#/#G;߿lێ;|AbHp%IoHlf7WWIpCkģe1j$*]5D{CƇQp bcBqڄQ@W d9ƞ j{ ]}3bm=T-muE0탛\ڿn]TU{ nz&-t$v3NY*. A;!= XbͦB+9s;]"P2`bdw# PTRmgEtR"e^,U+t'^63÷V{fN~2/^䫭Cwvgpv^׀?2 ;Z+ huW N4ΞijV֝[~S=׍ΏuiyȝW{8amdU\~f`t(mk='[O`1_sIԃXue@fn\n/0P߁̜$cqk$26$ݦ6O9WvH]AjbΫ1cGĮda5u2,w{-@qia ' Te6s'_ g$q`,#S22OC/WqZZv\o({K_9'O(^tiǗOP_8t F|͘OEt瘯n&`z#nxP1F4H aO#CMDk%~URy X"5ot]CqK(|$|fq-TP/ tOZ*-&&ż:!9xZzǍ}r3a X1d'SH| 8Ƃx ÚKFEQ¯Ӣ"Mڶ@Y̶w|r nfq'MMf >t;۝ 7?)lj3t~JW&hDFHtsؤ`2I"eT䌲{kS]y&4J#p3j!QPp@ Aw+NːCR4h NS MS0nA+%,Xg4:y)QؒhXp?)6[ PCuUZbGl'DEGhxNS .aֻZ%k1$QqV^[>Y*@S_t5XtkmNRIF☑`EXθ" ؄XȤeE^ͣωGֵ >WFށzFQ1t'A[#XN/4 ,Ax$LY9OX6jk#^WZ-OP{LJHZ IkW_% Ty u…V' ku'u ߿tW0qhB՚tվX qWVs$$FxilKtV!ЮQ*Gsg=~|{w,/))9ӞSE11]6 T&X%FY$:`鱶6aY* ||~ᔫ\5l.-~[~Zݳ{I0kr]7yCD !Kldhښ(LS%JQ*d yl5f1? [ϴn㐼L60ʍ͔mqg?ͶĄshm}|(ƋъQ> o_Fru+׿P| = E֯>oa D7[uJ,&b1Y㩟}>#+?Lq]i%#$ޒQu{]39KT]og#?tG7]EhR؝ n4$z⏧Ǿ49Vb# (fY G@ei͉t][o[ɑ+ u aw<Ev 7KFTDbVS%Qslf$_d<,WUU_KBԔG؋+3L5_͔C|\R lt0fmL<|Pkw&V8־షuARάʛⓓOW{pYÀlsNmS2t`RpM,؄-j͸X] -צ3>~E|@Ή<8L ͶS:S VPT/: CVlLmloc_E:!; [Й )PY2z6maO`xS/c /GF;7Nh! l)^Ġ\(늵 Jm}k8JVs$n~{yT.XCF?C_[hUII͸RQDȉCS%FAGPyNVԜ$:U pS~D=&f.Ŧ$gmƚPOKЌ꜄"mV%12Zr2᪾^agHX=z{#\W?ݧb\~{}PrYIJbd⫢ny"ϑ0a()SYVzd$&ϐ0fjD\'0CF jӎU.7LӭRa'm?:_+(dm& #;^|}2ei`Tdd PoI3 9FΏyʕ$~pɊ 3>)3$},}Zk[u Pj2악99N!a8|b $8+2+@2W ,&eHHތ)X0) !lN(J7)vnb Cu90-\叟6KgTş/NӺKͬ]὘تLIRq6*B$ⶩU9F$fx3ɡՍ|Rg9XbʡTlUh&:ah_{_J[%ٯY빰ED,߾SJmΛK󋣲8M%-V __EN/^ )MKx\L7[ԗ7?9 v|rͿye soane{ }alC{3o5"P:ը iqn<mGnӛjՋ{N^nk |2#o?4I%]mʾJwE~(f%艛 7i&۷?.];]Mȼ!Lzc7iS  <5='j8~&ϔyƻ؋. ]֩?>CWxk=vR7b;ٳ/e_]vЍO|U,j_qx:4q2`nW`_Mj'-+ ~?[ܳ׾x_K1?1kU-!)6 UK)&y #v`b,Ϫ?UM97/:QNzѣ&4:( {yيw-h C0T]?pZֶXY/ԲSvUf O##aD!rAͬ Tr%8Jh |h5_S>kiي5@I^Eq;yYs$}]GRgsd&٘hD9wڛHX¸>XQ%i}hԉss$P!UcʅT!T99v/M~c|m$i4i<[~jQ?M]լ#t;Yof=[O'xj%9A֐9 imUH,iU5;Y]o}hQQ',pф.VUszJZ}s T-dB>c49e}~/?{ p w-:F2Y'p'StH_dm/k0D k,_ 1XM!qQ3l ȯLY'J}mEqvw-o1M7wq=+>E (Ka~jVO"P $qЌ]ؐpV-mK(^٥M2]W39_,.JI~ɩ?]hh]@E'8M!kID k-ր#$'R 2 RneDK"SFCN%rSg%K>dJ{!Nƻb#5!Yi'ZXm̦Rњ 2վjwѥ/KGqZx=6/,FzݿܖRk ZaE6Uj֒P0Tɽ Bl9){ |.ʈTM2G$6kTjeݓT<1Djy-<dh&7-hkTj14@Pj5%.u'I,ɔk˭T˜\Ɔ1+^-ZB\ԜW:U$ʑ(QgNK7%9/z/aI hSJJ,z,-hn/W8S8$^jߣm)Llş%^[c+e)Qىyo Du~jgm_W/BiU3[ɫ;uuqpfzqdM[WG}'w+fek ЫfA7j^?A{OP8ӤTzPćho,V繾>v[zB/ohzx5/ևkW^K.뫪 ԲtCտְO|p^wѳkMi疼$dMa9{B_FK7IZt>.zvŜ^-~:WٱY^Bb^,>kc:&zI+m"ܽ-}EoOi{~nH{|T@C%j$,1lHِjJ)ƜZi.)ځ¾1]f(z/Qi0*H4Ӛ#$S,KLJK,Ek}Q1lEiVNWkxK R\FN[R- €Ra'-PapjJ*-!9j3I* -KJ-)1hгBpi2D{a$7RǼwk¢џV5I!q0(q15mh gRYX%շ7ٹe 'MY)k_wax}g^4 \VazglŌL"6-#`hFs%l;WӬʉTX|7]3|H9|.0 $- T)JA[mW N(9Ws]R vvZnWȽ' & lp[y`0^x#x+8Tj-&x#32 j0D,v/ʡpe䦟[ >:M폝|zn`0:OE2*0G 6&@tY$h) Ӂi]ICE-{GY1 3* D9.# dxqRxV!!: H[XA`Ir,a#ֱ<вpZi]rMsXL>x-ZU7qyqq,Lq+^&\aoST~rHš_G9Wj$nw5F-rɤ[߯҆tF>%0KMg[ߧ\qy#߅37pfa˜mdYvx)䖞ͥmsȸ^>>&#!L&)v^$&g=~G7II$5 ggXw,}CeA]d3Wn 4!- 4k?yXZ< w3Ee&80wQ֏8vB%eq%M_)\Cc[MPԤxcL'F__G݀k +rW_=_uL/#K")Bd?ҩm(P 䗏?8v N|3ހ1T)b+t@HmOjR0O y*p깇4`*68] __߇&Ǡ6$}7Mᾤ5^o۰ȵp޸ma8:\M,kvdHkˁX{3L0Ds[׏1 ^\{@‚L0L:Âo@ 88kO8k/`9VMOgzVRUN)`Qfx Z]."0  iBAWu Z:m,FWNUZ Յ ODl `]¸b!@~2iQh 5T(V8d.>:.1[PsL(QmGs iEPZ)"V"D4NPvjtvNzbq!;vYwljv@MFМ늧Kl▱գ 3`?KYC*J!+*/ yt h$*8ʝ7\HBSϘ8P=x:FY^1Tk%i 񰾤;8ZܮP&yoRaϫϯHoly-4|zzB ,:nt;?EJy ;֡GYz^,C"S\w8o7pjK#)v]+9fr ٳGe4v-Wɝ_50ύC'^^9!7Mv|&ٸ-EpDK[B]x[p1,R,:惣^J O+Sh坴R/QszWj^KrM6`}>X@|_bREEFbK<<˗p JVAZ_j[W0 SRQE*Mt!11%sF45#:,j=%l)2->[B*nܯF /Q‰r ǁ 6QRg FE2-AU(t".s(joFI*SN6^_yA)ıH1 ފ.ORhww7YB-{A5L8~>]X=L 884A[0cqͯ*>Z<9+\(< F # kkU1PR4 dp"Z-Chey+HXG"\JlB O]c*5( VN\KKȳ̼m8-V_B))(F$"FzfN TAƀlD~RBb ɀ8㊳=`^DpjD6d78. åh2XbÓLElH%1QMKc5@5ʣ&j} ]]rDfl$(V&O0 .&`d"׊]ϳDfFZHEF#mh+8 *ǁ9D3/P(1LorR QGTQJmڪU m2;˜,jBHo~;|*K4bɮF^OJ#/+ۂ.`)%or+rL [dQ(  PJ"{gKttG&?=~I]oE<OF.ukF]9  鷻޻ދ/V++钊z5s>w=?3%L7ZOExgY>(Qx[xGz졓S+a8C-%7.騟/4J ( Sߦ5}>y1wgp)y}DL'L ;Pjo<Ÿ !ƛ?08r_b燿oO$Hc׮J#ۼWC=Ll>;Gkwg6%"?S`_r2:eb={-fh!ZjhVjSJ R߰-5DK RC-5DK W)B%u-5DK RC-5D i _CԸ%օv|2-}J+潩uCBF6nw9~AHA$cS#0qY.c< FJ(f%kT6o0bQG"`"RSFDi*a$EV 1̾liy7ƥxo$4ۧq|/]OmZrzY*L!1x|\QaRjE/Shg 7 ;=oXH8a7Cc3Z`(3*h2D;b0XJBSELH\quIeFg7_%EwW6ŬvX,%u6cO6ՖwIm>琽8NRe<HgahWޔ2Oذq0OfS֮\WlIG.5d"4 Bt<)']?3$5Iaz9N?`.0 $- T)JA[MoZ<֙WkB}MB))&[)Q[M*Ffd jb&f{QZrJoIÖx''nAyg7Tf0}M[ upP42aD%$@tY$h) Ӂi]ICE-{GY1 3 r] L{iL{qRxV!!: H[XA`Ir,aKIsؖCi ui7a0Ezރ,0c-ZU7qyq,W` `谜C<@\їaT^i\ӏgp;AWNV1YҕD3qD.>V@%^=s^qj0z|Lٻ6% %G`yI$'ȇ!)Hl+oȡH 5gZ:[飻zo}Ĕҏt5￸I>^@n<^/Oȧ۱5Ϊ ;3⎀h˴huaeїcu3:)2̜,k囓,Yyd(ݚҵDc+Q"!%K9 Q Ƒ&&)AqtApLɀRuk+X J&1C}m^5fP5}'Q˗×n9sg2sigrh _]ZxfXA+扂S IoP;InԓTjӰc7w B'ҶMȸUU|w¸r$L୤[ŷnl/8!)2fW=L]Ǽ}M*u2WzsP7PbÍtBkN KU$fR{CҴ2(cު3*DQI=vÚ&jɴ,TS֩M}튾ة;X5|Q(jzǼ N~|ep 2a!ZehV[&srν/އ^So~`W}=t4m6΅e5g*jS pjCRR D0BN*l^hE\fWs-Z~! YWٰ?.zPXP}sLr3/ ;q=ƭ(҇7udta~0gm8n4ih 5(O+LDrc#y Бr 8M(rnAGEdTw9lVJH?B"I ˄±cp 8 3poF̲lݏ;"@wy<]]̀@j)[('zSeM~8?֤~`4{Na2R=/W2:ݣirSHaavrs̫l3oFꄯ v9\) 0֤\_Gz]hP՛c$V{Rr]B/?Vs'AfdPv5Ǿ8~A0`Bm8I_M硦|z,2#].bnBreFVv;]eKIeY̙C:j`gn&V!kf e{?fz/?og5m-4|k :5p0oM̏T hi:SM1WTʜ(}"^#Q]ßwЫñ7+u=g0ցn~Q 9kf>JgG-f~Y$҅S՛R,7B (AE2هP>wyR1e=.}C fCے$Wx~v4Q;)VR )ͧ5Nח-"{A'KhB݄qTC ;&VvGk܇T( %L835"&~7 ^ G|1jtQk'a7Za,4pfc7//$5t? VN/HU[D,T#; {F"r>@7RP܌FvGcPsw硫o%P0 dGtC(X`R\UauwȎH:ɉDb7ب~Jtx(z` UQ(y\K(*N574υeрdi]4TnJ6dv#. c3 4f?Eu;qLgG..+nü7a5Qbog z"wӻ!,Ō0-pTr]`TcB'9%zYx\GBBeĢD(~(ĥR2qihfRRr] 0P cJge\_ kHՃyĊ05{N/v;7R=9Zq!@M'zBu-_]]3Y<F/d`$dG6;w)\*A[ պ19mI 'Rc>YɱTk*: \gp Ps]ݛ?~7߽D]_o`10v:\M\j1ӚpI$a'@ǣ'yC+FX wLJr˚шV g^na|k,>Ugٜ{}qkgbxmſgeЌn4x _&dޏ1n*MS*4z}E o6Է=9c}HQNZV`=/dv`DLt*P9Rh ]k`2ҪaW}9t$+R=N`pVV?g?=n^0Pz-o;0ޚ{> !!؊ 4g rC%̌Vj dDP&ӈ,$LE}UH@R#`b:&^r Qs$tw߅EWn|y'>AzX>!n}6R1he+9åBza/c)&ʅFOˉvl၄$ Yd.WbVHKăë쭳(c/_ꎜH˗ôw%Ͳɐ$5jxT@"7 qfA%*Ű[)r9Wh6m~+g:÷[]vP͝m<['Ɇ5d>Z7x#ũ|p).}ԖU`̙,X{-ʡkOt-Fmv5盇48@vok7OAeFѰ5` hd0D'ZK(I fh/#Et$LS%T9ր5E˾"E.qVq̀2 A|@N@kC䅌{iLs"iA QzHL@:łKs@g HXӲ H{;(RO zg~s_Ӫz.oegH;hgOߧE/OEԢ{ w9 ƍ"k }o:10A~tM,QzlmM׳&;H2vC66})-,SZVpSpsE*LRryPDpC%K9 Q ƑI d 𝞭L /50w[SqXnso nGW[jy_n{.<\6 s2 0 05˺ WGV3bk7FWSH﫽rJw A*57mi0L606\fe;Un^VMmkoIɷ0$c6|_ĿebjU[FԪj5ޡٸj=i'[ <6 Ꙡn Bb깘yTfR#sF:not`KuDvRcW;sþKU8'|o#o) +r/G^E|!k3ߏ$.GJK' azTH;CblRVP[)}"^#Q㇌'yQ;hhs[ Y%ƾ92?pyzV<*):VSaxXi}iF3oA f(bcV2,OdHXG( >bs@IKTntSjk jxq,7HitL꩘̰4N+D@'1(܋0N#}ݱJб =wFdkT7R .P 5 QQxXzG!uV(_ߎ]QMϥ[@G xQ1,6Z@}0+1@P%DF)@ pl:yw%P8˽elFa0 <.Ta w?x`o#?wI>5{#rfW40 [5sjP_Cԃ˥"7⛆1_5; B#L9FgRP|wf:\Y=E,t^V@BRr Q匰p-L}j>~JRƳ4|:1>Sw9M3c2?p>^, M C tZ-7=焪;A?n@W\kOwT&?Vm=7OnFgͅw7l`w|e q4-ws(QvK3^L(d݉߉w 鿍Ͳ| 04k`Ť'm7=WٶSmk}w{Vis!tI7 >L]+GdpoK S2yg /A>?~~x߿w?}o?}h? ̂p7w!ixܻm[+pT}n ,7ÑXܷ1hn[Qqͷwca.IINwZa"]i_5r24wrWYBp>Oi7n纳Y1Hla{ {-ѩjvNb8(C,gxɢ5*Cz Y=Jz\Tknb+c-+SR,nFYm&ix0vI~߇m ߆In9V{w븇orx&}6CI:u*#=#zՓQfmQ[R2㺺Ec`zu0sHW[]ES1beP+X/f}MLryå O=cF@tHbaJ |"앸WZѯDýP=L׫kmٝ/T8TO1=[y)?La309 X0J`D$4=-exTFhOf}O85[M H)HoFAC?oZp'A)MvdXGh2XJf@3\/j8Ow}V5MMc`$[n|y(bہbW6bɾ ݾ_c#>`E]1K\*xAxP,́Jx*-#fk,;Zh/,6 e:B Z>2!G S*ibX w B*Z`0c`j^ˈ) Dk45Ddن*j99a]\gmVz#MgY>7l쎮[-דYӧ{tg=AnڼU]umڝ`𿋉m//U-Iy5k+;nM~rZЁr BCD(Dy`"G^ gp(joFI*a"mH d r"*DL(.e < dlv%8m ?UEZlMqtٽӗPS0a.`lGлw Ae%X`T9F$㑅H>HԔA b FрG!eLDھ}0$癄o.]e]ڃ#A52 cU#jj']zJMv92ͱ%Lw|Ne;Un^VZoqOp{[8_ kDX3};vp=~i!Yg <63]wg_G>Eg޶RŤ#*-)4줰Dzt]b@y5,.;o&Ie#*B,xřX5Hс`o}:RŨ+edI"XavsHeog3h ҝF]]7D{NufHN~:5K!.G (v]ʡ*,s餒83ၱXiTX:`)$Ťޗǭ ^-b1p4s%#⽵$ Q][o7+lݖY !kg `ž0")z<#k$'3#.sՒFJ+"MHvuU+-I&u[MފV@JPY/$jȪBUHt|#z`XV>d/c=6Ŭ_j,Yp=ifЄQu4:=MG|ƒƨFc&(gXDnnSd8s"qǬ"Ԥ^%yotcYY_|&WMv^RcW^ixp\!gkD眢=hAwTI CsEi;+O:{8#f+RYǤNZoNgKS-&cݤZ8/lPOTSAIeM ݤ]-g;V㺯Yٮu'O=m|r|v4R%sXg31OY#ܘҎԆM"RS]REWR*1 43@VȄƥL!',e$' VCaA]e+yl'lEb{&lJsğЄhh8=u<F摞=dAc&e(EaJZYEt,ӥE%^5{t=x7s3 :r?*\\̾׳ˬכKw#lˉ[={uΩ6˿6j%{A\ I  BV# 6^k&%.qaMii>Yz6B26W^` D^K2ށNX]FeTSX3٫`wF(sr)ER1ᴊIǀR&۔e]19du-a|wQE]ɐZ>ۧǴJ#ST*`$wh-\1{>6 ʑQ ǤE)xLkd&CsJcH“/:tZ8(Eع(H:sF1u5?$AA%"l\q'w[yl XoX*c 6o$C1g{P2e \c!F(~) IzߚjkFD&LXM`TgT"FylV Pt RUe{zv1g19$Y:qQL מ pfG-wlSz"rOq.8&|Sg`1ЇzSQz36ڻ~䲸Ҟ=.kCanZ2Wz5}SB%?pV1i?LcOOX ~#̯-V䠙 y~}$D!׬5byFK}ty.;ϓr8ޞԽGU$s*1e_-~[.勓Gs+@ 0>muOG3F*`yJݻg=^_+?;=9zݰq gUKEj_NOׇa|NQBɬIgjzzefy˂x(`b'قӋ9'aRuf],`ڝ:>:aЧS_ePO_> &35=V~͠y߳ o͏o{~w߿}߿ow0N2Uoت%,ï]¿?et55ϷZfNzN-G^1u1žC:[[R?7)E:cN]a]U] -bgQ]?JK{*Z!:zn)m6e4q;2|U8LJo*ÅqF_ vW9ƾ) I? 1kư7I *CoWϓy,m/$e(Qk2L<<<)g6Lo Z\R )#~:M+_v]R:P_r :#AmXИ;qBT?x+;v[]\Q;_p_uk zx++[--z9Jnp9d*;/}sa$A+IeiFL6hl;P&oDlbGK"j2EP] 1+6o]E]tBr ]G}^R1j׻gr=gC]a\Ov@;vy|qQ[tz~8L֌>♷:v]b_iҥ^uΚ_ܺm .d:VXM\< 9Rd<Y Kɀw$z\)G$H*=Z]&iMڋ*7HP(N$T^DͫYJ3}pNEdKş޼X2V7U1Bcx,/D{ôW{&38~nfQϥR %[zK y8+qXס ^Hu"uvz3\;.dLT&`j@mYñb.rko oNe粊%֠ޅջzV]X wa.ޅջzV]X wa= Jէs}\>קs}\>קs}\> Th=EBjQ[6z+FL*yhFDȪI䢳YL[1?AɀL18,VBVPl1YY1+$XPr=r6cD% Kh}fiSxyh??9 j^=|i t]t2mXy[t iF/}n 4ks3Op;gv 3=;[m|^ݧ|F񊡡@B M|(NQXbDBKQ{tHA)W}Hú@30@LXkOqxNBxlƋGѐΖF)hei,ҙj3oM*:#ĊZ&śтe%zHaє_'mV@^RvGRD ZeZ -s:gFagp(55C^,6hz*a)`F,j'(#KxuakmHK,FK@{ }xL IV翟E!))lq=UU]]E\ D3ɰl*(P){3jF7}b0AxM$m$ $+B.( ְ0 ,o1vDI#e~YcvmUvۚ׮] j݁r2&!qI_ O wM"UVwrXM3nD?d+i-GVmq֋][ui Ob21-..zidzeL7'[$ dvjveXmSL`4ey Rf.!c mү~ J cPnrw~py~uc !8;]/d,'?ҍx˟o_:, 9~( p6{[Ά){qٻP#Pn;~KFͨ!Gp "@ *3 C9cDID}YW #FHG4 e8čb\5T:RDqlh<frGȍ\$rrg<BLLh-+3\܇gc#8WfZsD S'zPA@u $ TiX;[ZR IƎhY]85l[]/]x':{8rM.n}zt3͡Ճ?s;?]$ 4,k4.$LP -1@f`"|L+]&[rĉJ , wZP@si4< "ED  Iтl1bahJT: $(GC $eTG"*rWy( i-Zi'=`N8j0"pÄHfmpmS?6 fmS7$6@'j} ϳB\ av%԰/B3?fm Oyv|xQH=_'Mgu:F~ؽ[bc>2yO+o=Rĉ.kq(qBYM^[,pO#abj҃Owlܣ\xӔjqwlp\[6偁ՂA6?WZ.L "UbKN=hI-Mns<dն˘tx 8n1i%xV6;H*i6 5 b].[kz4]Nt?̻Mg6KU=wiz7i]{CaxO'樔DL$02/={r0-whq?r™kW+/c1{5:/0K ^xR CJ/:~P(;#s9䑸 )R#U: PS,T!"E DEF$Qtx OHt68pu\] 78 ]az^CS*G痁 1tl?מM֊d:J8~[<՝#1ޜa?9.lrU&VPk\+.re[*UK;43=]Ooي΀|it=-7c(BQz=o.l>ijI4buKMͰfTs3bP(}Oow?o~xOqG`} kڅׁnjͯ7EY\6M}oѮa%_Dۭ ڮ58~;9,mF7nWMteu0 va~Y[yOQ˽&o{/CB4r݈@_mލ{Phcsh>bgye:U RPg~hV~ `Cط ))g4ᣎ6xbD Zsa5:vt`p)}F5h͍SДA;ŚcAi VR ">ޑ'sUCcy21Q R)Z&9b"q.-NBv&Q%+&;&l1c:lEM.'KvM5o ZzVTsx[P5}ˆ E`)N86ZSR0+X2m#ELwiOY"5mNmsj6][f)#X͚ŶU9̮aR*Qi>2 *@ Dσ$WrcyrJFYh)TyN 7G*n ?..;qYHm$/!AmȒgݒeY-eɞ fEX/Vc[Rv#]2{fE,gh(%ﭥF&Ab5s>uYҎ fQ1XS\o1j讓U%^5WiU7tt$H#.b3W[7dLWgx;}Yv;O.rK>$]joKK-# ܗZrV:H0ˎ1VRg1m31wS|={/.nzA'tNb&ioοY~3[R'z=<N\c5'_kԱXM|n$(IoY] TN\mxh}-E`tKsxmO0xR *sDH]7cE7DxmƫE_+`~P ]S6](ÔBnK!j yے64 [ >1$] ]޻_&^\ß,byoJE/}pHȨæwa?}ω!aOcqY.#;R:6HC,\Ky}RK**Y+냉KM ` Xy$RD3e5.|J~ dӮ'3soZ^$8zms;]{=?2\iQaRjE/dR >3|7X,sDʛPUc tʌ ;`GT KIw JP58)x \I]Y^,}Q{k:tsF}|>'h)7J K72?u.DӒ)KC%/Vj }$߁J^B`4Sjlq?1Pb (#0IƄ-N(^`#XiἢY9P2zE)PG S\z ^A ƜL*7gJ)G0cG^p^^P  Mo|5HjBuW9zǟRλWOP420fD%$FtY$B&LEt!"'Ul&(E.qVq@13P  D9R!|a#T䠰G˞z{E4FHHubq!YobA% 9ē +s)upSZOiMa ޙ_(Bp]ZՋp'v\Ljj`6GE0G D [\( krDL2D8ANV2YҥD3VFF$rY 21ĥ}^jL|1ҏE>^~[ޯ-0g8:_۠rtn,XwKQ.T]V\",nz9IE*LkhzC+kho64Ulκ]-$;}L>B;p& J~~T.A<'9Ojԅ^̯bꃪUB2T{ܵYb7A.\0ז,ku‡C2[qFiwN3ygw'P/EcDRA*0aC{NjNcTW7;Ư[34u#Zœ^=!^*F`x##n&/h<80cI 1dg%"3X%gH^fIq-AN8VcV @/]Q.# ۫ue).̷\khC  Wox4`C/1# ,`%UEH߿%RDe3!U[Nl %@4dtr3ϴ;kH)EԒ)E@ׄާz]aK#Y'Te6,D <>P0vYR;GjdrVk/;^s Q * Q]A SJ-W}}P ljTzTL]>C-e÷ak>.K3][Qؖ`_e˲C] O-_>6y#, `V;|p3MhU!NaR_eo(ppdKk,)QV8dm,{X vtqsfY%V]p(RLh+1u6䨈$:ϩiXz$sBh !+r mGs+:2,"Z)"VD h̝!8]q\脜-2>9o3Ԝ|ƺfm6F{%:raD]m4^zNzSz ^(\(wp)BSϘ?@1b"iA4ZI} Չq]ܫLz*_2>7?ϒ!|T`-8{QH.8  ؖpNCKK)FD_AsKxHx]T?sf][M H)A‘,7AC?oZp'A)jzSOO+1`;gVYr9sO7ޝs۬* ϛS|m& ++NLp$J* xF$x;om@ 񷳩M/W 6HBVȷ~sˮ\yrrx Fj[P%ĕ (s1)p6-XWkH(h)6Xkv6Ȅ$1LA^ c-|VPrٻ޶dW ~ dvffAW[Yr$9spHYM[0od.D!eB- k3@ f,`Hk1LFS9@[KDf| Mwm ҐaQL@m\XxoϭaKٵE)Zk⋲τ mjz^i)m-5AVeKuŵI@b5;2ńW ҥŢur*5Bxs]%9WZn_dKպA#H]wі WëVzKrҬn5NsQGOgwhSIjnn~ݜQ]^ _??jdD~Q).@ӂ) C%/V1Ya `|1\H0XpMl{,MGA,*rB;%0rX0} y2T"-)0oo.*[NBU7 uٌqq>+_,zz/~{nӛ!xc.?yJode=C慂+[Үb#9bRc6(SXR~ )&?>A3{IH;v4siRB#s&hG>fu-gŁius6 { ԸC-V7(QaeV`{(*(us6WT`]PY @ʖ`g> b6!S(<_*:lzxԟdGzJD2=:b+T1yN9R:#@&*Z*~"Gˆ^Fc@,DDꥦTH0<)c"ҙc6r:Mqt m}~=~ö#>4CXs3\Y?': [匐R+Moqct(VY9i#(MF1ZleF͝QF@#*%( j۬(dFN;?0Ŧ {'_Ͷ3ušM^k,e]k-OU(} ~JNGl SwqS`Pݧ՛ap}V;$7#>%_7S؛'n> (_^KN>,rf^d~J9)440)g((3F)TXŹt|ʐ$"jxT~8!;FʩR ˽)(49,. Xr.Pb*:ٱԪ7Gt=yt?7e] cnX1b 苰ƄWJ%a^`#)òfB}DpB))& zSNq ,x#0b`̙ rKlK( 0Q! e^eZ&Y47;Fng: zx 3M>7.Û74$xaITH'K!.Dʝ99N~E<muƽlQ-c+6@u4H>q D?X.#ȁ19lT'myu #(B3u6$HLdu4(@mJjh̥$9x紞zN'4)28ß,;+YIc&5|XM,שfQr*)/`ijBG S`%UT}wꣴ})5NF#OFhrqS~f.UA }K?hҷAI>w"v)Izv{Ex2>nNKܿlz*%)0+盞oo1F`odA3ͩ[4.;d\~%ւ=)J@ϗ%9g1fsh_zxf3 V[3\dOZ/9AU=T>< ԩJegwcު2[,MGcᖺGPieND@'=} ]|:߾tɽPN E1o&Iwǝ1˜K+$ bW uYp֧[,Uyi|\KgqK¦UGs)$׽a\9Z寻^azgepERBQ) F-p1edBƣ]鈎N˧&\\Y]}(ؗd9:d3jF sl5Gfq-1iTJ&i`qXhoH& (Wm%6!`BL1jveI4a8,3FNG:?|X(v(Vfꚪշ~x-ӭnѩZz47.:%rPxhמ@ h9L+' sYkG[x BF7rCֹAؠszGDJNc2HfES%1#AyD|= ,l`~z(f%e =wFnS\X2 QQxXzG!u/,[smM] E-@G "*cB*!^Yl6!@T@a ;JHyi'3gwf$w7tn.^Pua6dYၽB$y\2%E&a4$+5Cjavm\d ɥSڞ"/y E`CLlv(8F"rv.0fʎnfsWga^S   քKx*0'1f0KKp;6 f fej)~:_]UP3ש YУqN[(˔nJ+wgZ͓,u\f~Y]x  )z7WpGËR}7+"Xj?n _4%Ci$<t4 iF6X#02 f0b'm=ѳ1WSqTZ?dӨM*O娳tjIsX>L]f+EdqpONҩ5Eg /a9߾_N߾z߾D v ̃ᬉw+ |O{CC+kh*eO]Mxq5 grRZ2|y5o3 n':횠Jz l~UjħSURMwT+_b7 d9=Ԗ[uu48!M̖T1qQz;eI'5}c`3A;GJQ&?6fvdr71~Jш28Lh4)FPR!ؘ 73CQ|oyvH-տ=dm3OR9րb !0L{h XsB9:*lkbG/ S?(+3W'Vad_1Lh V @$V̧ףK{ɱ$Q -9+Rp&>f$eZ}Yj)yr^lJa^%o ^ wE%A *&Բ(0A?2ܩȗ݆hOUǃ9Kg>hl0X4Lr)೛zӫ_?' %I/@j0xψ<ڗ3rA@ŋjNl%~괕0J m&yZlŒ`^Fј_fP sZD)Z0la% ,=zZF.uvN~ j| BnE m%FD9^3-8Ѓtᄀh !oI14ѣM(rneZGEd2w@X+\$J"RHDclۛ6YNw-H[ >,1}6fjCX"7pTv0BmI5KUBqXtBE,!]Xnu OˈAh0'5.5,e}S01 ZGK!zƌg>OLj9+h4DSM,g(eZ0֗ЪtX,FjV \v)[{)^->M~^рm*X0B`D$4={ d]Oedg A@;(DGOYV{g dtr\tG JK|S9q]Yȇjw1٫ߔkBZ~2>W;uu2kXv}XTtG鉅5n=: RHR?яYw Oӯ Qq!Zx':{o$? 9֓H*bT' H-6YaynC7=!i=<dZwY鋪j@?{Ƒ MpvT !kg :aU"L Iّ?áDR%HtuO]RjMqm2Ċ1P+3gE E#g8{qs%Q9BC eҢ616)Ӂ > B3]' xUcl93O@I 1jM.W/a9F blNjϗA_gg5^SBj4S" HDMsb94Rh qJU29'+\O]I1ued0,`FA+%,F"w4рD bZ#BiN9|F'fZ%ka1$UQcR\ko]Ƣ3(.^{o8$E |0 ,ddlg.tokM.X2w rdc\{k]5^.̷M|RAMxۿƣր _pxM5CZ 7>kbWϻͬz'[u/~/-9<\HM9[3ױO~ywKU2B,@e6L̄V:i{߯/#޽Ó/^ygI1%{[tQ4EsΉ#*4QDkiQ0}/!K$7}jv+샛M^'Lի˻7'hK*"@;ZY%7:Dv[@7oN(ͭ$^jIХFw9 sl0Zd)D8R?}4\oȃr\|\ՆYx3L*? ;w]% Юwf|%ٴDC2V YDT.j[9 mT.'o Pm~s BDxnA+C{ޝ*/Vʶ\qGؓl]"pTwrQV{o5Is :@q:4GoAS7'G$ dgcr-]o뒦N{+2S_X'!JVS&FG^$ e  &kd4Jc,o5B(NRN[-Ug0OIm;u7מ|uJw)f<γ|"(]vs>C]>:0%Pne.p[nyDkPLrXa{VxIPƱE[<5yZ-+g_aіտ_Rd *I>%j⾞?o _ɨ7}"I'XœD8n)e6/L" o|Q#ѿFHG4ͿhDQ 5PDpJKo|"gƳq߅&4W>nuOjFuF{XN -UF劗5.WUZyY:%*ő5 ƔrHdOhe*FyQ(J!$H,S'zPAb) [J؄Jf,Fg;7'х8cO]ش/ Ӆ҅KW9~2ivqٓo~//d5<D$ sҨ@xch1#^%%X#-ۤk!60˲Ry NBP#-F&JTۍa<];EkNkwv?CQ}PwZP@!= uM"(t Iѩʖ.Fv}X;&Y1F,jDVX#N#v2X(SB\p^CQRţ_9%-m:TIZ'iG/iZ [=g jS>QH4dQޥCv)}c{wZYn,+ĕ09 g<=!$ ^׋[%;YY9 knqr:oYO|ڵ^{sq.TMϹ`x֝ի'N(Y(6eeһZf'mDˤ}@1=,dmK,MwN[ wn67a]0fiԊP]Nʆ |,nqo1'rqOۆ-\p33nԤsxZk;{oIyv9+e2c=Kq`7 47̀#;e҂D@'}5+7v;)IŢwSɩ{g Mf"p"Q'/z,`-1'X^f,l:z]u9+Bνzݥ?Q=ruaT9gWRQa\4\ w"T>Yi{!w؋&7`PsΣTaOJ>rv=O`e#[-Q:åIPmB zoya`[Gȩu4D @'4\Qc'awL.DeB(y @D%,hG1Vn% 8ZN ^aRp)Ʉ=йL Fg קWcў,\YF_pz8&}S) 3wr72RǞ_yu#LM{NFg rN/陒ߟ$^dJ4L\WK- g%oyB1׏ _p 34Op:lg{Խ{V$*ʾWӋ˥8oΐ-i>ww ?Ԅ`\28y<R5[Sswov{3u_+?~7]gkoA|ܮBYmUVóvx8|$G~aX0}EfypT(\ْBOn_&i먌lu>ɶQfI\s d>?j+rD=lfC%=߿D>2F)(o~hV9}y6ض1]BH]# \xe:])m|4Fr^`hJ֠bMDW4Al)µ]tѣ#s\բG"SS * j;kZ洊wi w"vPe6*OEu/glE}f5je9j & x=d~ڧ `6H@t|9HLPz"P%Z3ĵ]5f=gUwjQ\SRWɓeZ5 :{ji.*`wItOZ#ȬP\7`oxoWP+t{QV!Px__PX^o۸ȵZW?{7 ? G_-%Ӿ./w@f2oxzEOX28T2 C-@ًjNkhv2z"H%燃Tɫ43ۤ\0 odŴX*gHŸNztjw^9 srC0PRjh)ʽKx.K># уDtO#wcBPL1$SI Huh}b)?SRJ;,FxӪ:-5[q1/R3a`hnk$v牠9qIjZk珚>I y: Y4$R-,"T'r™*1;1'2 ,s1(A +5 "tJT@A!1ZCLM,m#Y-C\Ъ|Z~,Af\;X=%9[z)~Z}@x1 턮 Z^9l%)a F=Gwq֙wh)B wLHUX!^(i-$3)p(vlNx?3+dy|[SpO_E™K#@|$T;4RRD7E8JK Q\&[›Nxo^"ofS;-oy:ߴ2^BKkm<c<s]nGBn+#3&X,`cyٗA^\S$֚͋&[MEMXqNV#|zcߏonG㊳.V:BiQ. ig\/R,Kٲ< }.VݙJ¸r1kguUL޹TnU|OGj궒O9}N+ԝ4P2]ސWe<3ucj%竎ʺϰaF_bzE+YA̿8/bk|yq ޶t kE*mx6zLfW+'gunx^Mz=8Woɏ/Og{wr1*'7 ϯ?c,_Ǚ3(k;hNEn7·Bx!DX[oK^;-m1;6@]4Ӵo6B*]Yyc>p8|y=GߚkJa>b] 1;1XJ2F`jTu&#;rlg \d8ʞ?ڿ}h|a7w5S:=/iL/^qsux\[s99"uW#гUɚ;5RvRR0٨Sqy\~GSf ͨ:p )cT8>\uFw+qTz7xvMVܤ%u\[HnIkrpm0G{wywgg*>#EG(D"0a҈*tk8=J}|* b G:fDyF9YrX\3F>O͹x9kQUȫ;;]SI"ֳj(%%QVĠsHr=RNZQ >„}^طbX2Hû);R3B0QQ5~/4FIx 4gmGKho5Eb1UW)6QO}VHad1X(R.Yc gIQ(!Q!S#PkCu:v }J2l#_36Xv)4}FmQxv(*6 :j;q9#`5GeaD9qn!VˮWbIbltpiEFS&@ ` .1eC;%0ZΦ$v ԁ`:Zx _3<@[vQh -zt-ق8(H"cR8fWcycQm4Bg$@ccGZ *bV(mCŖ\79=|e7>p~_|_MiOUu$;T n=\l$Gf=)SZ[IlQfq׶5Hr%'on0FvS%xbi؈MŒؓ˜wàDlu@ck1B[uwŹfA*WRR ;h BxK̨IO#P X ַld؎lF R>]Ow؊QDqmbdL cSMMZQybRK$쮮zUzUp%\u W]Up%\u W]Up%\u W]Up%\u W]Up%\u W]Up%\u W]UpjՓ ۯcͯyM.rCۅ*=#׆%9\O~+|~@ 0փRJGgW{ ].eWAq~PJtZXǮzz0gîlUvIo9QvPL:v5+A$] }]#$4\|-K)8wdvz5C #`m`PN!/z~z'US9ͩ&=q I$XpykjaO'A9QT{0?J $_Tr{67́@7o.™_ek(p%S2:͛gMLW|XLO6OYBS`ed<=y1SuS"<Q<.QhW|Fvhv1uO}R}0sHg[E|XW{mxMPVe@9c;kEE>Ӓ)E :p>k9%&&?"Φ#(6cFLwnw1y8/JɅ%l,b%h PC y# đN*!2"wR.j tGr=k `y8}FM?aAZEw=#{n <@C2eSi+%߅"u+:cѲL2BH㪘ޮXІy|:][ia,z!3޴oշ}_vT뒲iφ./Q‰)-h" pm1(t\+,(HXw,0K=28 )Ҿ9r qUl$aP) mubCAѷ Z_]XJ?p.=LnJIL=lBuRg8~x" 4t^ͻXwVP+"S:m>VEaA K#+0X'2R Z*QEZi Ҡtr:aDxM"b$\y&oDng` !Dd@qJ92( 贠e Oi*}Jʗ\e̴ЎN ?j~ʂ~Y9Ԕ PV6|S=5Œ؂oldsS$oA$ơ 0C'>L z'phfOyћ{iGz_g .Mw;JW/7KVûEW.9Z 䞣>CVFî6mT^5YS>W\PY~;;՝vus?܆(7|mW%_ yoȼөUaK.߮S>ϗIUKF܌vkl^I蕋+O)Ms{T"q!s4fϠlz~ Jo]Y0TɅB7ǯ%ڷuk-UfaiZѴUMK miZ>/274HuT*惣^L OT38̻t=LTt*Uď: }c}BǺF*@"FE#.+(Nb:1*SԾ^p) D (Z=jʑϟZg8tC>S\[ Ub6{8jg_͎>@J||t*TL@/̛\>ȸ=,eLb=*iTX>2t,Ť>[RRR#_]3{WaKBE h(]81"ާ42 {%%'lWV)iX4Y4,QzvV;z耿@Ⱦ8w5-pmv˞FiGGҎ(w !<Jso0lwւzdBr B}Bd:!!N=ssDsAHA]Z uP'-UG̝w#tGP $]ST]ͱK8Z[g5.7}9?@2$xaIbTH'K`!.Dj;-psl;qy*\58;i0#HIA3-Mf@oZHEBJxew0g ,d 1+ Dn$Dg (x`#FuK>Us}HT8bkq(G$-sDqĎ#n {5rD0Bm8XGyHL|z ",iP΁%,M0>Ka[g;G\9/LuJְPZ拴_ N(K?dM+Yw%=y>,/On}/Wf+x/"|xn_gmgз .AdnsSL_iВPk/OMqzS$%^MG7'I7XHQ KKI RDpCJ,s$#mM|j]\nc2h`-6w"xlTa>lQtty:tYe̎0;<aX6 #W&s~4QcԀһJ6NۀK6vaZ=lIڶM׍v/b>IN˭;Nn@ W[#)*nVxa7,LXx c{D8Wh;4ˋ-}>nr3!" TG Bn8;6~7{LE& MOq`6 4e{NҒN@tסWr3Mߏu/J*#ޖ'-FjPP%DD]B:>rswwtnΕ,y\pN_Z$K , b$f%b(FƅK>[SM{@')ׁފ1_E.c 9SD.FF .LQ;;x(&.@gٿJ``Vt! r.aUX.g }w]eE!$1{Sw),KiEfZ 0GB-`5h &)c*i2-c110 f1crHxJH ECCTWhAd`k.PHx3Ӟ{:FY^10@Cѷ* UL+(3XԟIO"Y)U }6~_\*V8[#hkPonRN|}~l@x!B_*cRb2[ܷoHEj{ּx)% R$1xЂ;(DGOY7`VIMΫ"(M4f}i᫧Yȧo ݥa\@XY!vB`1[-%QRiXaZ(H$HRMmd9]@eor&l]3 9;wm7BF.`^2 90r n0i—DlA(yvTęKEΎ~'oM£j0݌`NNX69O9j3j Q[i[7RLư/r@YjB^J,?șzg噣B<:zQS ƒtܾ p`rqx-΋Te b'R»rN7nگţ\y)*?-v |3G=:)"P'@ٻe 'XpBj$ }O5o}uӃ'? #BfE1K\AF( { tmb6T-kH(hɩ Z;}dBJ=7*@L<UXlYb0c,S Dk45$nhXm}X ,Qmn^]$+2$O^.7*®Y?ޯPTHEz<7Rg$ʖ쒪SxNOw\HZJ0Q]$-"I bIgDYQae{ɩl^Mɫj)x= Uo+ kjtIހݿ-L,ͪUU['X5GZZϽχg=e(C,Jn*q7D)1T@0`T\EUhUJZUwJTi])î Jn,؜]%*9nwȮTDbW@0'`U"CaWZ]%*lճaW|ũHaգB]=\nc%hWQTb]G+޲u Jt5Rp觝]]D/&duQNO:?EQك)&ǂ!F1'xVJ~@lfX Nq(l:Q辳Dh˦6M4z@ &pPe"WCaW]]1܊S\#{V_Y_ΉNOOW'ߔJqotV*MV?5 UӅ2 oସ#$ /PPz>;Es,޼__g0E|ec[c;({3΃_dp{k׷-]5Cꛑ.w}r#F1~ѴUwsV ZjɺV IEjj9,C{qgP=cRT*`MUTV%,Wo^^^o.^~w8ݛpF`\"PGKŸk7xSMC{UTUn倯naTkڽ}x{1ooC/UwR*{ <Ч&H>a_MLQ]YENwbDLrW>g! 0z@N/:݅>&-=%2Bixw1[S FivJ3T`lqk~Qw1CH D #(wRkC3L?F@܈1a8FAN1:@ l!*[ȓkG*[> H5/dm3OL"q{Ddd KLR 2$`AM{ %(o7\c#9ZY-g"n ($6pK xm;Nam#/=JI=t*50]7d*0j 8 Dbxx݋ ai/9vf1|%gSՄކO~I%ڏ<;7z[cR;ĬW逰 ߜ <0XefHñ9Gu.1:gQBA*M1何ևt0[Z{+QItzoq%[+IzQfzAg>l"]9-2͘A6ƒ EiCV@ރi6$k{|NQVbl$QItS5ӂc!A 9!i6ZBhT=N9DŽD6ȹQjEubrX+/B"Nq9w]BPmv<ʼ^!ł #-gh^o>]:VRۍ13>pP)sQ)nq})np0sHg[E‰uXTva.F$*8ʝ7\ 3f4=tHbaZ+Io5jkW3ЍWyleC[?F>{/:rf+-#R2577)'>; } 9 `hf)1H-h[ Oj,u=k^eaI@8P A\h@cQ Q}Ǟ "k/%%@Dp)"$kZ9]F~/5y;_!vsЄ9p:^᫧Yȧo)BtX#VVH(*@LVKIT֥fXI* wݼl= M.`l}XlAos/~7ݾ&v}r!GQ5k5xn N ot;y16 -:I bx&P:yS(zhI2Kbv+պz^+/MWy0 W]|4*k$Ղ;LfOTM*"J|ۻYu*j`x2O)~;~nozrH%wH;\gzI_!NJy ̃6 1i׋<%)MRYwdH(JTz(2YQE{Z *la b*16SnH| 8ƂxXN< @Iag9Y Ov-ZBBS#BE&m$X-kedl5O{}TtoKy6xpL#\N @F4 #ƘC-wB9Ҧ)yU'M3[gܴ*mϜ`(“cYQƦm-#YZJ [ZUސL*g$q`,#S22OC[MnO51:1 F:5*$3K \vPLIbJ$%T[*oBRz3"u9&epr5ї3iڼZO"-,-AQjLG!,qgF D eS|I;29M'M_GMG(s/\ٻ}š8u}m"mUݨ[X֑" %hwP2K+om+ EbaXf2ׇL|루n /9T#IY@`3[`.z,UzsY‰>KRI8'9(,xNgR.V)mHYunSѥ:L{TcbDmbAƩLZN,X[)#g|hߝ|3P$rE\)NT<:W}!nɬƿ.Whwg{(*oStKD\1`kM1B6j$J2E>hVV)y!1Bz^  ,bn %DpJ#c1r(J9 =n^ kCa)\h%Ǿ~Ȳ|_{# ,8@IJ# o I:<8bK$|$e2e 36T^r'yPÀȍL>&.}9 \ jSAmҢvkxfV'c$cLj5q ,d8NG1aٳiA8x`^gQք,n4R B(`a]H"`NuT#blʩx*X?EDZi-"MÖ#<<՜),!rRS%<4HV:*F 8 ΈI) keDO:{|rO/]9[pq[צC)ʗ@ J=>Y^2#A2QIj$~UYYJwPop}qO%fO_R&7s3οfpVxy;ij:3s>[j 'qO ?wz5+d֮ҡ }=. xVnZ9I!'&}\ńs/PQCLPQĨ8<hXOKMV, |gEVVٖw~ v`-Z%Z0be0>E Lb>{4{ =nt䲲[\"`y<\[ ]!WWGV(^X(3}2QcŴ^%NVI6n300DҶ%0.t=%4qҢ)BEer[W]S|qaIi.";Xi_0c›P#5=h0?pc7R :QróbQ{Cb0&l'sVx&zc~_s:oX]bg)k*5u&uZ_Z_th:xP@yUVԸ}p8-8)}ryp> IEwbDJ S"N`4gh"lxqr7<4bq%ʚM'U's)mqR3j lZper8\@-%[pe*m[p\3f9>6U8\ej8uT+>+2!95WZAO2$\G}D~f" 0[w~<*F3#bi&{I!AY-;e_/:?~&{ߙP.~ܭ;ew}w TnwuB')IBVPS^Z*[׹xvHo%zvif+UO6.wŤvx%R4'=6 gP"ANE-mͬuB H y\t>osj"B{^54Q?aŁ ݩPqرު6m6[!z5uAFQ7;dkDUlko>tnSaō$C#Eo|b#d,`f /l3nGI#^y'es}' 7mckʠ/t> z?;{P3pzWf)+h_M5_;lVW(:톏SNku6+gT,SyZ1&f_2m0 lߥbSNFX5并8 F)Ydž2U)>ހLx~׏s B{z9ɕI1HdlH(<\Xw@$t[076la;K4_jw\&]y{-%դD>SBw+8;,ܣ6LqhGِ|3Y ,K|ZߞDM!*55+f8sc SaϏj1e.n%fDc*.Y:!+Yf͂3ܤ μVP >XbK lXs 滭X쿢bC7m$d;%CB3/pT 4B33Гf*no0ciA q[HYy ZwRDyz74揝`2`5{9!/\2 #Lo?omt#%t|5Q\{gl ͽ#qqvB&?U߱1_;~<`jN3&vȨI ƍ4r v5N Lp }Qcg;u/ GfkPis$[&i*->ewG9nw|źzUxOSjւ3O~ HEZ>2l @'X%9.9DDG8^?o 55KPph KN%f<Ytǂ>OHρpE sA6LJSid_b+#,C\;,SF'f08m7ፂZ KJtɦaZ1QJl(#b/j |Faȏ¦W.,5M@B;UR#!'>ډ³qbs&C߻n\eʊyBj8-r ETiۅ^Օ^jԚ`5ie*ivg{V/̋?hC%. R~71 :?S]b5)4ˉX\"3[ܫ0<EqB~@;1Wr_xԑI"|TI'Olk.1M 9wޏ/bLnkіOA:-3EKy=?{WVnЛs^[`-z6i %Wu~;&E4VH3iS͝ɡHйle^ 5SBI1!Դ@g{Q4Cwq6-=j:ާ`}et}"$:<(GE<2T!HI|mBP/aSǪWLk-oFzASoP>A)qRAf.%nL >烫q umTﮦ{51mt/yN3k gKȊG)E/AL3њ "Gp`pVym<5jx]>53;rp`I(a]^F3ħ*eJB]ɌjdI( Ie^AyQIrcPQ*zcۻ-lMnZRr2R%E{|Ftwy@IrymGOm}ي7wVgHs4yt.\a"-,iQ3Z@CDQU3j49=+5b802!RĜIPeRR5c5rvk(Yta58PJ b KPTm~U2!%],M2>e(I*kjֈ=jOzq,|ոP(+E^/nxAN!0Y}*XS*jn,d-^/.jq>T ɜl.5r7Elw\d"ׇC>S8$^RBBy9 (^Jʪ{}BBZ>eJqiO檇mN٫ 7$L'Q~7OBmBD 0,' Z.#]0 )X$cn#:FyhJ>FC\6?^W8lX u' *(]76})N5dznQEm1sG)vfR(i,P)KFxPxyƽb3(Ƹ٤PRT6b 3e"=JDF,t%jfb)#[hኄSp'`<(b<; .ʊ9{:$iUU t,~eu49+mUb{7Eܻ)r-75q?O*jj5OLS*2Cu3$@,P)H2 B2`V-i`P31#S=yuXoj0 H(,IBʖxlOxŌTf<0*|G76)g/^ꍁEA%6t&2oB!kM/դֳ }ޅ2/ E7ږ&"+J]Ѕf6 񇖑x-5~oWie֖[<(Em\p>{do1#\ ףP\0\NƳ}mxwb AˆLH*ڻcvΨMOx=0قMpq2 p,lfީ H!r9s4m;vDʠb43Bp )[#g%砼"h=NgǶ)^f%^@wjSVrT 2<pl!x gM|c(EQG i/=i]`xWg)S_>ӡTaPOy$HOYS|)/RgnmPH.|A&řtz3PpQ@yh5:OcƱ2V>3-q a,TV6 +yPր.X]!)ʙ, ZfҬn#>| s44qRxAz*xkLZ2:S=bE!QXP2 -44tU_]ޝl{WR2RNߛ. T*}QIeIYHFA1#HB:}ЄլPWun-E7:qcARGNf%I/C B"U$HD> U["=Kip!C=1* qu B I:E:Wi,b7hXі=? Ϯǡ_ji\ʓLh&UADFLjKGuJɬa HϹt3Jݒb~;ϚE l6_ Gqjᆿ'k8G,mF4&)/}`ޅQ٥ ?lvL[y&Ȑy?gL׬+'d @0~rNΌ4fOfsR'ido,9IKޥRLz6$7kazmg /S8}n=uҵd9K4?Zj~8_\D-8с9瓰W+r[S ei%-j9G;l=X4/nFukq›l1<_߇</V۹]s*||܍o&w?JƺFR#Iyax0{ufyGx(2e%G׫ݎ9ηB稜muu]v+ 47CZHXm2 <(S8d7+jSM?T^߽7??~~Ci)pE3<=WC4jhrr1⣘5hfmmH_~3B^퇶?ev$xX@WQ8$ ,h}$׷Tت$ q TviON96]4m1Fj1[S$tN2\d+㭖;Vn[m}[)JTZ=8tgcaN;-1)2$$r+aho`F& rµKl"I$>͡EbvO=IgT% DX8'#9 ڨF u0Hy[s f4;O&8V4MiFΞUwuqqQM @r<[=4|͍K_i.};*]snZ)Z}MꇭM{9[P`J~9rk3 `Yql6r:y܅Wnd;bɥ*B"'t\;BsJbu*UmtX`FR^c]e2Xz ,ǔ@KCT.Ej0'pb̀߹fБg=}9?KE2O Xg(s=a~|[ Y)v:%-Le)R~@brٲL 4@N>m:^zu2oCfaQH.iaDva8]_7nvK!E.K'`-WP\sɽZKAr#U) @"+$י ͝!Yйler]͜ gޅdvy?(zCÜ⺻׽O>!Fa/-n0zv OShU&Bsڤ@J:ƬJ tda;P؍04`x``MbEDT6e{`tZH("MՑ~v3oVWuNL>|ڠAW_mQ)E4\/xeR|?Ǚ ckV][z'L̕o4)&4ꍓԲ!3l2ҁW/jkZr 7t z?tJzrƩ$eC,' ]JTk'c4[$H]RE#90,l.HX05!g6gkJ^W뉋REOb/^vo+=S00vFq$];>n-葇Q\Ď!ס70m2Mi5m]ՆKXH Qxg'I* *YrƘb&VY[XJJ-OCeZf UGC;;4oQ1=/cp|TRمDߥ3iB6&ڄx֬rU2MdKuU!+`N *hȩr`N|YW}V^5iǻl=p^W"KY;Ïv84ng.ۓl9Ҳc&w=tNYX6?] (əw&|-ʒ'. Ir bv9!R(~ue^>(GzRyoA< r+ Gfi㹮LGa@y'dRksK3|>XuBN,"IQb6LBt(h:%W+eJB5CɌj M Ie[LI 裁*zckY`/-MRv?\йp~[Jɤ&PRI]0RD4Nf5Eh& |H p"KjșX^`SP2n1B .mV2 kpU˜폏sWPZԆڍ)$ '34i b*5G|847tL:i0"aIHO2"V3g="~U{Bfj:⢬bŻzD@[!0E}*8=UX -1"n1;*G G\<;2ʇ lvُkg_2KN(Ip-Qƶ:@09?8W_CA*B6Iu.̝ v^yeýIز!>F]7i{oc=dGz"ݹW&5KZe1BthLGt6J /% JfQZZ [#s\I`WQU anc\,_BԆh4(dz\F8!z49XƄz2?\45r}cW?=|th+Zfd؜RsEW}e:8X=&=ѳ~~fݕu<.%8t5 c]VN`oTy2fTBET"&bKW2:t1wTB.E`5f츜jGpS tYv Җ*P^l{B21σx+SYژcAY2`uI$(ݠmtt!fu7fEFp΢N 6HСB%/6r %cjlX;TB؛bahh*ٰQʪ4T&sEkrGwF4qNQ{k Z(JA;TCDQzN0=Rk(^.ndcF'njM( aɌ$ơ0ɑ#$-T#z0w(( r&gm * lˉfAt0Wi,b7hXі\ϏBjhđXʓOh&UADFLjKRnmSMa%KEz (͑:҃6r6c.&~I~}f\L4S-_(+reF Nb)Y"$%s`ޅ -s0ɿ{7\gL{0< d~\ &׼(}@ @0a鋟O 짩sQ$,X.ae5b\%\]ko[9+6;ع!^Lz`6Hɯd˶^v(ھ݈aKYb*WAxe4׳(}i=u>g,k57ׯfs Nr퇰W+rGS Y(3%zJ+Gn=Ӣټyq3ҟw~nxyvx*9[#Dq3(s+ܮ8h~ .'zz3b0pGlˇ4g%zi9HK)+Z>OB++C{8V:vKĿ F~wo7?{߿~~CΟhf)ӰM"|'4rhr>b򚏼a꣘5hzomH~ǛQ`E˰:d *p=gqb^|d0= .Z~W*T"!G6*qJ q V6x2|]VT3 XvN .Odݟ#%%L&|$ Q`,FgM/J#m9xHVAh!ggNq6y4hcviL#b9wdMߺx)h}{d0wFsܣIQ2MߥԒR?sթj#/ֽdlv=||c ֝.qXk8`oux+@`q>haz]oQM٠%Q$&ul4 M7RXd9Z˥S}d󈬯KԍxE'`(#<~v.}eUpVSQRC PIwX[dJS)^%݋%_`Iw_}p9+{qCqp+{d9n8hceu`^{o5A븯t:~ g΂Й0JhCƎ&c(hMBކU ERdY)_G,AgF@ҁ!s F&2&n]RG~H[ʼNi0ZIVJ{K4P(ӱԬ OR+ |*ZveKHMo@KjO3+}9, }oW`w2^lzyֆY)9}P04k ˶BS{wњ=%]E(O{ ԉaa_ VäԲ[@k~  ev2W"x2pUUpU4WZi}cfޫn|g',>[叭~}cq2hP%^3q;ooJ6]җj>3¼ fo> ;Syj퓯p"oildrF:Ĭ#a{3OV E1Fe$eədtLiU,F`IsDQԭ(FfC,U39V"`1tJpH<JN 0 0q2V)FriOo4ZNW mzC{/}1ѴRd!1@ B8 )(oh`%N3BgS*^y(9.Ӵ'7T]mnf:boI7%n Z;kDaqdlyh^hJxZe,a;wraDYiqMFYlE]}[|Yj# TQ3XR 7\l;^w]Tݍ[E^Ho.]!Bۻ.~$ pԺͼi:G^ّzf!l岬n{=4sG+-x3<>k,ܜ?Ɋ6ƳnsL%YUl-6?m.O)$zeO'񩄈izHҲ>DuKMf W*f A9*el,g]g U)VuXGvqϮXIgWP%JXcBȊEKkȢ1l3”E]!D`9N_ܖ" ʥȡo8}, |f"^EiQZ,R&Ee^ɤV*y<ݧ t\G3Y@#!o$Y[ ߹).4}Hv!]f\}">Dۗ]NK4NX^ͶgZ'bZd6qD̪|!5tWo()h,FFI41otG)U{Yh"UaYt6-JeYLY3'h] aȴL>)i(` %?|/4 }[׆Fq*I) L{A:mds .)J j "KC֔%"#byN,nu/ȩW{lyfA6CrOTaس# {Q ;,.ސ;/O=65&!Mwι:6)! o3D\b%^,9cLp1erܗz^D I*e䲈 u&4I Vf3n]h.'?>ܯ.'}lIBmG[<;YIpLm^918mE>*ZJ[!i#oMg^2Md\a*8t H1'VpT90W*T-qvkZ g%ow~W[ߖcM1ٮ-Yo>HzN99Ml9ҲaJF `1 Ў~ RnRu͠;doPIqbP࢐+'(fZ,!Y9hب+~ Y;kg]*4ݑZ3ΟTT&˄HGd JgD6)/,E J٬PP2c9LSҀlCRIIJbPQ*zcg٭ä=*"Z&wS~~n x_w(ZNG†uZ3˜G<*Q3ݴs45\ʹYDD[|YG haEUdh2'*G,$ "LXht@62VgBոJ5,2bPXx+sa34,NMOA`:{戭@DdR()=40RD4Nf5\X6!A:E4 s3iX^`S2nd1B .mV2 k_1%nQ\11VGWPzLpV2CȘFO/ "^s[HȈCyCo'$t $Y AD25OI<'YEȨN>&n.!t3bK^Mɇݪ:(ݭj%w[jtyFW2|j́'[):br!f0+B)\Thg<z6BO<@M!r.Ȓ.)l sgKp|dz,f$Rr<0,ÝZ[s8^9Aee=</ (J8m"qZәtLK˼e" d AY6WNOҷzZl}G P.Dj[v[@jXBVK5@(.f:øVLEM-H]YoH+_&הk_!d &iKV["IEɒ,Z舌iݱHk9u֭{5#'c@`B-Bo3b\/=38`tw<:W7xdL縉:#C&UG{Uf&mgxoFNRuf]SA]w̗{nzb+Ds%02)h!8Gw+r$hBP}Yb*úh)7Fu"N3.`*x/ ;6%"QjIEcpP׉Ģelȸ v8E)V˜K ,xřX5Hс`o} Rh e^u[67ִY]z9ԍR:z|ODOgH4~k1" ?cs5WTʜjs]Lќ(\y dԝ ߲kSIUKP/U+ŋՏ)Jtyv \<۫qN(70W\Ӌ[)U&?fkTݝoOfF4{~«UUq S?q /uKmݮID>|}eKP]I$ƷR_/fxRr`4Oƣo.n~{9kK%R/YWjݻcrYjb9tL.=Y&;<*='?l8_ 7M߽\zu/ fA ̂aPa SwxSECyb[WC45>|GR[.hn9ҏ7__aᒗķ~h6ngMXJt=_|`0<7,2W=DMT[T/)D8LمߦݸtZܻFZW9wٿlmْ'1f)p/rn_m= IA?>@z4H)ʝy~j&uF#f܈1ǡ8Fabt!^*F`x##:r/1[)̓Sb5hg%"3X%gH@[fIqwPlB;>qwVzau򘭨[U;V/h4VRKIZ41TNFe` X>^b~XKy$"kY"k5aqxywO]jioע|t.ѡOR,`k U c*y#lp6HF_jRz\TknAbl5A&)`*._7#_ hg_8}i6C6Ym໛z]˸F$ GY8y wu/2jU~)ΖjEF:mFKTKI*M&6,u"7#l4W (O9-r͙A6ƒEmC^@!V3sHA;;'j߬lcÍH1H"*Mi1ZL> 4҇!&${4A1%cTEǸ/'qـ10 f1grHxNL WC?9cuװ+ud`k.PHx3Ӟc:FY^10HC&#ML=06է C2R<zv-FV \;;. u'OxEW'T=OxBd' gR6u 0]]RLV.e&n`n<ePEqx_(%Zv?\""Ο %=>A\`.;DWJp ]%Jzzt%fs+̙ ]%3tX }Փ+%Rѕ!Jp ]%*'HWZ0I s֝`+pW*}l}<]%ZtЕڱ !:9]~p9;-]Vh1ʶٮtz:鱀vedmٺvՕ*Ջ'IE 4o0A@4??_M'jsS}HfD*4]14Xt WuZNXhPOOa4z3Qif 3vuPQs/G"u2 8꽔Nuc^L~\$kxw^F[|0Eުស,FXZ t7ǯo^kFɅ⾑eLrJ͙"<4;˜~~i1ZL> DJX!AVCDGЏaNX5,5J]01ĵ r B($<Ai|1#,RFJ{ӑ&4]IkyMvPAdkb䪪m KI=n^hE%+*er ^#QV+]aL]F·]6Hi |QyglլyYrl5Gfq-1i.paf4jD‚DVKM)PqZFe)A@F"G`IY6az?oŲ׽PB2t9LPb\\p^D+5O~·Y֞hn`]>cuJ:夡`ўz\j+5 w!)3#9$PJ_r7)1rPEב H)ZdNF03,*͸e)FD00< pDzCA-}K7 VN=Xwsϝг5h (b(7^"q{-b}?IOQ{MM c%فTgv.mf%omlb U !`3{3HVs*,"aMN!i&ɞZ`( _$L}5,2t ;`!1̊.<`dB#L\* LQ9#_Z &eb}QO[7Iv1Ůu!ǟхkIK?.൒=_/D@ dz]kMniJq72I_.8#$MNnTy&tn|xq3ҟ_/gg5H7|ppz6[aovQD6r:?Gئ晤Z?iii#1bLq+h.|<^-6zz3|0۶yY9>bfݴVbyviH )+"_ƓAW>*cxG9Vj&9h( .Έ燓S/m&CQqeZ]Pi=a BΓ3V4eې5$*Yc&Lx ٖp!{Ў=Q҃p\@ʀ6jόϥVQ{om9fYJg1 SQ3\UuD H*sdlJӴ>GȀe5!wV86:b5t} @wlGƓ:jjw&e$Oom={ Ӊ,ɖ1tw z|1ka[=ɭMsjEh;3NEۜJݮn棷Mw#mf F64u۶wU4r;=L-O{nּ͏Gy?MiWw.#pVQyM/]4c3iEOC;`K`wҡX˖[oma5|%(W懰\zWJk~5_5HXXJr_ј"oildrF~KVJUs=zei.z.ȍ1(#KL(9s)m ,i49 Y՛FfCҤ"D,܂N)"2`<(b<; ..UV 4a%=ܕu8ڔ5NHi.I\ [7637U#/E И-ѥBj-tv홷 v[]Rմ~d\8p RwAepAlK;yd!>j%I0e8hNdkQשEMYЇ->"omH6ťܚ'{>"빪yWUR12Z=MÛ xJDAl#65LR0!V)r7WdU'H5d ׎uو9FPZ !XD)Xe4.o,M!1%D vrJQq/VCgHv0~uFUlgY&=X^J:k:tָJa?rXf]CP$$'3ZD hr$ .mdILIk4ۉ㨘9eB#)2hs]VVUjN,WZ{-K\^s d#I& t܂BluL$K 3҉&i^=@5]qv핊Fۺ_ ݁/B˧kx3Qi5W(<-f!|#jM q9./TUB֫oxWq]]@L HNom Ad(*&&<]w {{ kAtЮ=l{M4-cpF\E^۪Ƿ>YA4]"Di1pqH Qx'\p+CܜK>H,9cL ;T1=$|ݻ).Qk)0HJe1,BrIj$OELef+MUCgOĻ\L^?:/&w.a9j:zl>f)9 O.|FiRBK[!i#oMgFbXP1guZ. 0Y:K"1'p*!Q Vz^RRn~jVgM1ڳٮ-,-y. 7lͫLl/ %sJeÔDF5b.Ɏ%OcPJvTfPJ}8poPD+m+'(fB#XB%"s:-iݭ]ww> ݴcpꛩvzt4Q 98irL$}JF0 E#3"͔QqR6M'If,/ /Ad(w巘KbÅW0TfjftZ 6RKvChRNXeQZʃZk׫}}*OXC>5@M$3hťDȣs)S>"Bf58:DU%FI}rt`802!RĜK-.-GnS5],FBzYp%heYE~z5 ݭ͟hp8c8޸V F2I()=ii`h̥"ł5` ꖧ(m(ƞ O"6Zz-6L#iMRf`%-xLCըP6R`Q<0idF)c=1D4I#x847 )ѿt0;3N :9|zz<Ɗ䴄,̆jd_ǹZZm#"uG)n8s8(eDł&myȪnlc&| 73{酣5{a5gk続yk c S2?JDv争8OGo<6[£pV*q wqKGP=ވ M kݑT*k;J6\B_@p0mPJss{܎)ESΕTec=1 =>[: Mw\t+]JnF-7lP7j7yFc~A,gtN&8O+]mhav M;u}KoA߻3vnaj7;FJePKL1SȵO u rv'T3\~:Ɠ!<"v0K/_#ɪҼ 'L[Qn ~K{emFEOwt"dq>$|:ھXɞٙ+Z~[lӶdwI4VwuSEVՓ j|'~IkXV^IygldF(es7Ctog}M猾{̘}Ð__E% C=,t8z*s'Kx-?V7eSlzlO?DcՒG_au@Ne7g| &?2Rq>ga؛ 0E k|~!|qy>Ç^"ѼjhbYiP@Y3b'B&u &[cϧ<<{G?Y1%+e-g70L)%UA)sⳋ'ЉK@^?Q`HyRࡨ-JsJ*EL{-,}к_J+q'^As/՛3jm؍|ۻcibn]6N32J{̱ 8sU\Ui3R039>ŞX`+UW}*-]*#\bL$IpuO '/*2{J)w]i=\-!C\UZҺ+R5[+Uܥ_ ]]S5CztPn׾|%VOxbq87UYM??~|v<5ۜ  ?# ]M%I;?;*zu,I _g#&TSwz0`gqqvr-Xac6h1bw6?{<_nϿ} 2r::^uAIt3 )@%'iEm8Hf8)Fi#]I r"Mʉ0&kאհBjh3]7uYrbF B SEc@2ZɄ±^:@V3JG7oR^I4}f :WOyrao 6ga^cR1ocIZ/ >$)>X x0IcEa#O- QF_Pg9u|ZGz7vRwMθd,4*rJ*X5eD KILjm;.$4mC Zrb| )4S Mc~L-/Uwg󯧳u4h畠~`Yq5|-.F}_}(JᚕKjJ7N;E[6-uՇksR9mV솽+mިڿ^]ƕ/On?du+gޭ뭎n0:ff7\8qvt Tya݁z4obŴډ2kOݐ֮sNJYޯOI`؝#gCɃ").K-v F&b@T(!ƢE":Ӵo(.yU̼Lt>:!ȯ3B'$4|sCw:ǯWvx#4iuP%9gF:ݾoqz$MiC.  fX')6GS>~Xm{ $s%Bˣ*8nff+zxgGs\xNmzXt>@k)I91 jO%6,F=@v3Ųzb,_ȳB.ҡӾFԲ+kP M;t^YEGV"'r(QCzJ4P>HeXd>)䏒T|hEVL="Q_QCvڹwIv<>9>j2IhKtL24&bBtުbA9J lZ3xuR8î)[@q1B?kR2Bd)HTZ!SVPlfJ͆R<#smJ<\s"nr1PzTӾɚ#mZ+{ϸDcMDuJFK՚d>R7Y$"WUlo@D)2SIRQ zrH&ˉJaC }aj"Wi4X o§k+3^ӕhypԋ8't~:}/^9bkPI v20K*XPGuh\$":b}*meMŎ`0Y- F .3%"GZ$]q1wiǮ6=2ػ45Y5dP`G ky ;qd`;x MH訲eIdF$cC*:Nup3q6aԯV`DlED刈#"޹MxScT$TV/Nki1͜G&Q,%H&lSDtB0!a8kT%T )Dl2Ȟt4(Cl]78[4=<ԁqqj^KfZP\ƸF\q.WTǚZ+EA9Vq6{%du4:fDQ({#.+xL;1 aǽXU7~|ݏQ*ٟhU\kZ_kU)hk曵gOt_=s{?~y=(I.ʄ+:SCA IounSOjO;h'\E&PY+DB[썉|% *:ϞstNjcS$9.Bf}LÚhp‚luΧUmO﷞/aO}Fg\ eN.,Q1 \eMr6)]ӁOA"4PuyPF`x`BZYl FC*Q2fUE(I7Z%&V3h `(Sͩ+Rt"5F{vИSTx[.k&B CYCyL>m6^g}j%p;9ܭO-^9' <:dM`yJ%*NHD"0K&8oxf4BZBUd f`Gb.Q nƻe JB C:7. vuU.mqȱVo-D] "hIՂq&Ӆ  Uv A1YN[:XcġMB݋f!d#ī'ݼ!xUa\gz8_e=k;qIN'B_%Q$á(z.)qHIM噞KUU]]sNAT[N9#3 ,=q( G Q+k\P|q{Wлdn];/I~t&뿗5l#Ӕd` NLBͺ.VVkbbEeXW!T|dꛠ%K^+nEM+2~=kLWQ-U]=Nj,iQ*v5](ω.kȦ֐z䚭4b{:IԖpqS{пpE AHZRdX18TX(r88]p::8+<J c.$8)6TA9k"6bˑуa7B/@"LGAŅ;/8jE)EO,ᙂ%p+gH'̫>qFR\Q)sF͕>w1@ƣo?\ y7QbŽԻ"< |{{;ad[-7<]Gfq-1il.pa9(In]rWb&y$ ViǬQFYJy>D#K#HIuYt}ojg/%= #_ .__D+mBu%C˴haVV|BSN*sƵgD(z3ZyJ w!)<Y w"%c#PzGDJNc2\v2JaQZ^I%pD0AP8@2ZE ʏ傂{=7;#aekP7R .P (b2z,D#:e9,gYښ򓍄[ EbG=TƄTB|JX j% HQ-#mdj猠͠Kh!)8Ņc !K53{LO%_I6\F 1 fL܋d)՟&Mi{YEB|y|(Y?c@H)D0:>cSo y[  skBʽ%L 匰)J,0T^tt~|fA\ZN.M62 G FU󯺾RLF'ߔJحU l_5 U e~""Br0)S]:'3#4Ry_x}=g aގ7E5Tߎ팡$o.`:0xjk▮ښ!ff`DRFVNf6{Wɲ)*A[m:j[_ZH إ _1K|au6qOJ- XUF޽y]ٿw0Qgoװ F`l m$$ی{M+U^4Ul;u=kI]ni>p$ mM֒(x+\iCӟkVK+|IV]"UT}{*F-Uʗ"D tpNIr筧Mo2)l N3씕&a؜<6```3A;GJQ&/ZcS\0 o_Rg4".hF9'1 zTBT6fÍ GT9j9րb !siI&A=,ȹ9BusSa ð; ͬEU}ouj<[=LֱI=s>!bz9¨5`6;^r# $g\Kr\ 3b&AYgZԁ/J'eNQ=e[cR;ĬWa#A29HysQ `ʶⲇG+rݝ_/­ ]})m$}" Kg>h8濆A.|{wWojn~\-%ѾౕnahxєWu*_߱M &:-FKVzTc<jl&'$-"7}Ѭ?(O9-r͙A6ƒEiC@ա7ߤ@=>wp(RLh+1u6䨈$:ϩi '< F]ǙR(P&x92J#"r;[R.k%OƮo%vFj{])D(oҎחw>*Mi3̴ 5`oF )]\=`I3󊫇Q󈫇Q)wL\+|WN=ݪ;qVloU"W}W@bUrfW$D "?Y ;K9CfofcjHNXfOۯo^f*a4097<8SO?& aۯ77}@0yvTjtR()VboU"wP%Pˑuq Bq8ʹ*ވDV"]WzU $>2)>Ȍ,eIQrxu%;iuv? ԙTIOyD`9#:*H@x*2'chh*h =u8ҤAz}{q:8SF ,F[73g 9YF"h4 ̬V(`x@+wcp+?1"$> \7J]u\7r?Gi7n/~ԝm[31xV7 +jGR}gJ(w@cD'ac@`B-l>xkca,f8—|Fz[TGaWXygI0e,eH" ʼ̊hw% M|.;u r7i-g@lco6KⷵEKF]ԛs?Y4cﶃy!Š `hn)1[ms?O?Q8ZAkϺV`I.Yb$1qwy؁ JqAtu͡ 2q<ܭltØ]Eo6pIo _=uā\ŻF+XBXY!vB`Pq$R3̝a$&Wv¼ݻ/w'My=%ml@̡عhl v۩S0BsOrxk#5-(YrkAytS/ L;Ҟ;Zh/,6&SbqO Z>d=D SLwY%Mk;Ň2G k 30浌&Z%c|9kdM7eޖ4^n{3Nmt:\:E3k-xnNWp'bW)LذfWT]W=M:L(CZ[ȥCw\$ T* ńcBJOa8ŊJeVVYV%s(׫uj zo >\[晒x8\R_vy>A#[/|銊 WUbYzIl]7WnӦxMwl쌜ն&h﹛Yݱ-N\(<*1ZpYPX[&FV`! NDq,n NNA0"&1eN X JJn@6"bHEN H࠙ 3  ԻGdTaKY-EHExlH%1QMKc5 SR1_ #O% V8圉 "0H Qp)Mnؗ.\e̴Knpk_DNt\t8x LФoNd= +ٻ޶$Wr/ 3]`0>AʺȒFx߯d%YnrI$nVUWuW=0rR 6ڒUmw1~Ue6i7_,B1T)b+NGjd>9HysQ->c )pv_A.hvvڍ$Njmoi^,B4L*惣^R OK/-NZ)ub<>\8"`T"TQP`Q'R!m S`A0Q.d=G EέLcaYqGkZ ј;79@C9SvD8Weuz0d9|a֍M11%01_{y>Told輍ͣ{X'o=l; H7Q9#1b@\di2EN E`U; ykˆ^FcD2Y+냉KM A #(Xy$RD39|f+ɗzQ:#U]O1@5t|n,KXgCWpx,[=Md~_ՠz9X-, {mu VP1D!3JQID[6C^A1 D$2 gBQG)%qxT)&7S`@x{Ys4U XP 5Zw=)J m/ fOH`ĹG_m(NIFuA""'ȥ\%Ϝ +l~{өzzBX[I .`P:0څA+aK& vDK ӿ7:5֟P"&7`8^ 7ImV -Vif e;%de~dKJCŲcBC<ֆvS^sMg\@ȸ6c?jF@Dtkk7յfՏ$zf>A}3T<$Ho-[]pՈr[T}3lcw{,~pp0-]6b^L4[ -ȿ>ζSC][9hT+Mjr @ubwLLIhw 3+|:sutskm&m<1Im%ST倪A-&U\5lI;#sw`%eO杼kEQ%rs_>z -f\*T^ v+"lJDՒ"ô"_"}C=)!C7xN\t c.$8)6^q&(ց:g 1RDt ["Y(ā.ڛ~˷76v^;@,Jo5S|ed ^WaXRBx x\'ծv^(%| Wr>X,Qd,eJ+X: b#^<vM]cY~gxhf89 ZJhd Q1 ɠb_R( %e]N3M`G22ٗf&fˮnZsq쭣W U"%oq"R8Q+ɹW"NT*W"+ !e)ՙ;"j$7gC)%̦Ty02p_؃e` zd {juİ@j<[1wy3͵T*PѻZ*;k|j:˥z m@;$#hE_hcV9V[ρB*{´ qcڃt(Pf 1VYE BleF͝QF 5$Sp'ܥs_Y_~l2Ǯ{oVz>E+{0%!=ˍƌHʨ %2\8.X8%SJ^:"oC8GђߡL1_GN&JI) 5X&#:)&#QHYu2LwZ D佖豉`+Rd ilf#@r>Pa~m߶`AlN#Jr晊1\G:0#iqK5 ՜ȃ .sCR QC" $P[)rCr_PC P, vefibzw.系dإJڬ׷;xr^;v[ܛWTVKHx|gʇIAcĂ&)"ޘc JfХ+-W4fJF0EXP=)@1193Tn͘1Mf#u!yuȔ.|B*UzHqG6DsN[gw7[U=5>&a556s!b$Qaq5؄rB;%0rX0[W`ge2vRDZdֈ׈F u>b*:8Ҝi'RpyGQb.5Aqb%#1S,r!;N@Nl 5rn8auE?~|G&'I5YMFaէK/EǸ.d1tVbҥV+!ҝV'jX뵽ewouս.\&ƞseG縠 IUp;oB!gh՞{THbaNwOYm*{u(Ӷk]ئ8tT:.\Z+S jc(6 \>57OúK01=* =ܤ5Fcװ޷U tFd^6~s:1"`w+qkb*{^i[z9`M1$b=M1p7vL&ǟ~& ?|NmU?։|xz3dڃ2g̕MwW!^{|8jJr԰EeNA76A,Ơ];W18qxd7gIVc,'UހVŨ^@⪴_Y~Mas]mDJH ƒTfulyX#BmײAyˣ%)TTʒjK/]Lxps(|xGw(,<<䮺0SX:q7;^ f)ձ)pTB VSaxXi}iF3o fhJ~B{Of I"aTKM)0qZFe)A@DX/5%}l3R:edJXZaӴ ^du 9\H*:^FfɃN9itGr}ǣTjDIW~OPlT&|w"jxS*ԕT"!v#_;McF&zAp2)u(ɣSpv4?#o@} vOW"F>Ba'F׆r~d(}uhn$'y:5hXT`,h 0[iDC;r#3|jUvGMGNN,TRT2:A6o KR!Q]'\lNCl=zuA(𕞊7p!a91,$tY;z:%l%"j2Jk;W%Y%8 RT!5LXن<ГDpNQfr 0L8c>.cavڥTDAFcv<ƕa 14i O~S1q67Dccl+gp9j}ǖin#vE>D}8e®`u/6gwQO\߽u?iA<". gqR} .w{"!FfllH#<B|I岹>z m nljbڵ7=Ŷƻa!kAxlKJ D-"!b'Gb垭Q8s˿%͊ >:"X8Tec2*INw3wWxPφjQ_j(4?yWnU0N 3Ǎ= :/߾4_6tr=yA:E9Xc ÿwzCLԠ|ʁ#~4ƺG_}AˉlV+U8ȋ$ z&]=F$O TΏwoqj=4{M_]ቢIKH%M2 ^WWr*Y(nnͯYm*?x$;oڃRxn=c^GoMMPrmMJyC{٦Uu(va, IMsD}}+)&B<Z)x%WkU䞅6#f?c,Wzf!Q5`5$MJ$CAHI)R#.@`J QS1ek}D|,<{{\eOͧ4 eoi&&m&UjZɊ煒jN( H>9QfbK&zM[$aa9B%!&2r$*e"4AR .>I@y'el|'+(Ds jõ #cw&qIsgAg ]dL`Zf%CqM,'.RxU*UrF s22ߦ SK+ %"ƒZKH !$+?4#B?Sz3+ H|$VjpB>q̡#>". t6M2PmRܧ;ձu,:{SLQ4 Y*p(%.BX"uF*VǎH)_Nvism, jKnUSnWRMJ|(Dȇl{攵Dh{X.vPwʟz+BЌ/F)IN]h(a!8OB'6m29C]F<8խ8#:p-b 0uc/n溇콇/—7.-;z vD)eP]X$xl+7(@ IiI*xpp=Q:AԶk}W/*L."WH'D>@Z{R&3h(sd9@d1q .DxІ bVhG (ODLaXL-Fj8|D

BnvlFBuK7Bq}띕KZyX:VimS$οw$D% Q,)B2dRe6ra!-XBQ` EK5+nQy_٬Ky0G$"{պڮw_aS,ȲdR=|}1hlnՄd%RyDEG+˵ v@ mP ylPcP$Ky2d6K?E! 4'zz y_KQƨdb@51'(w[ "RHF'S6!WR1UgKr-dS:UG&Iݷ-UKE` !G"HZ yҊH IIQ4)  7 S2ǿ%hD%ьK r28 g٬lT!-uT즭bdǛ7JKܴV2Zm¡ ֌srRI_,rnCIyl0M-yX@Emv48SJ29Nk`ژRv&O Z`">9G@Z\ $NBOA`1@JQR*MR8[#*Ű /A(XxP,K6dHz;mOOMw~Ae{^wȬg "i{DI&0dtQD4@ ABqS`) ΃**yђΖ&3KM(&fD|q⸚*1u%(xN38'"H8=dPL4 2"f YA [-.NqXea< l~A_ǓxQw1cv(Npy? [᷑pz\V^EQ DeJ%fZ萠^1'Um7n{Hz}Pɓ:f&("=DIHTY)4mB71'7IۨCD ֋1u]0ӏ^׭0q>'?vxk79*Yޗ~Xx 4D E4i*,&K O"a!4nBfIGQAI }0\/q;pf/j *-@EʇHMJrrMf ( )iꝣQ)b%+/*&\pDŽ)/'CiLޟoV,;dm_-iTӊ\`iO2DQ"e _c9޸?$ ODF-I$R11 FBJiB8AQ)֋dcdoӨ%IH0$c,/dn}/ųY'^뵦^&i kǽs&r D C!+.>9~<;yղtBgkMzES7} ¶0Bݽz.ƹ^(D A!b?d:]}1f'~M;ŋ'[UfՃ88Zl:VuaVusyw آ Icw?g/**b>t1ٷusjUق%\UrGIk\5AV`^b1S3bv \k^e;9_R5^_mWT~=lf u(+ц;ݙ+sB*N7*x&ȴ6Ď^c;]xSz~|p~qYv}nxaMVPPK}Oj-(V)*\lCUo~A,~7@V$u<ú+%wheNja5iurеg}y?c,N/0Q4]1N׵坮.o16W_u)Z͊'mnVLKv3CU-?yvMEîJS3En5,*z,k@Pj5ydumz1qI& OўY2Pm,fi:>gM@a>75W58B<Ԑ9x=̄ӌK% KpE AHZRdhX18TX(rzhsgْC [$(eRsi$YL3A9k"T}8bˑуw{;L/QyMKh'NoW.{^3]QRƋ  .(>F2{rɀg ˿H6-$8&suΐLDFCҋ$[?#fXeXv2Ýlzq*Rmg #f{j% >1ɂi)Oo^}uy\|}v8ٛp 00 Nsk)m687?N߯?m CSYW_c\MڜrøGb6nvl-6ח};v/.GSn_wCq. ~ #*w @Lh-%ww瞋]-S :E֎0bkh!H 3A2wHsIM~l4֏`0;@_1~Jш28 Ozc4 x̆=uѭ#5΋>2ot32 xp}P)3Q캽¢s}=(rγEW8)S݂WAVHe?|w?OCJ:zd1 WcMk*do_% 5[j9:>Pb? vl{X8n1błfނXhN9dJ;X*Eխl=.*ߏ2aHy斦s&x (^ %Z^ju@[uGթx+Kג\ 0sHPº^fWosHg5*CzZ!lp6HF_jRz\Tkn7z7kPi(pw5?\ۻ$̊ľ/WM]:?R} 3T' H.|i˵|7ص݄muE-}A鵜kfԜ\haX[(=ڞ֦}`c{$X3^mm &&;af5I`'Qz}~Ղ aNșA6ƒEiCƲ2m@knt4l3-p(RLh+1u6䨈$:ϩi 4P?J<"VFudXDpaREbD hl;qkା˅);޼qJK'^%vJ6+)#g))+,T4ƨ*޵jmQR>l3▱$.[x LqY̙C:<Ncʠ_~HXגꔞM$*8ʝ7:113`OLj9+Pk%i6F Xw36I7aN/n3UBۼEy =Gỻ_Xl[Bls8'T΂#"_As#ٻ~d{Wڷwo71xRrɈf(HFm A\h@cQ U{[ xh_!mM/Uf6q{sr׏|?^Q2!G S*ibX ߪ|(CSHRVa BawL@RZFLh%e5pLQ濄~7 fvfjυ2^ynΪ|臧 rJ'6ܹ1o#\gN=*aIv:w,oI, JOӟ`R6@!pƆ`?.;IX%nfsyvU@ï^wd~l໽^ߛ@tƬ- Y*MhS6SeyA%T/޶$o_8V0#)6gF\tkw1r)[60:PlIHM{lzaw—S9ՋNFY#{4$-,mY|(+}qdmqF7Džsoblt)[. J@:N AA0Lӂ<ߛaGl`S2WyOU^h8/>DVt vϵŧ9R\IHS^j5h(#Ucwx(f]:=w:U576 D_^3l;r|hfeo58nSɱT9VGs1oAAP{3 t( NiC";.BsNI4T:,1sƵg$.d~X30qgQ*?C KT3 U2`mWuV6Za{DH=.@SqS\`"zsI aE OCp4S)[Aq=!t_$TD#^Yl@ j%f p6Ri2B"-څ8%Nl`gϽ*Q:u󥩅Hr\N0_ `ad?7Nf ?lX=yv 7=NA 䍛l_"Υ:f<׶wO ^wBOջ(skQ匰uuDE_ 13b,?lh FRa 0TWpx̛C |<`Ⱥ9PFΊ+8Bm(U@y0uV6O~0VySx9\UN煳a|c{yU[陸"3ϐ  xH?Vj;zoHWI(UM,tesk.SqWq?!I/ü^?ߕ|\K.coB0s>llVq}>SH ew=gH\%OsP&& +t{4)J]:ʑT$,kPu.&iqr`ښu0¶>c{3(wA$O: ʔ;.``&`KG{fD?c 6;}GfY|*"Ac=~@ :N(;K{{s3!piL?u¥ƹ*z7ƹ8vnۯ?Gr(Q7 F(pnG}${HUl8Q7 F(pnG ϱ:[[7zBvY$Vy9n`nO%7ۿ,o;/pfm@n)zH#``:BEg}NCVqygƛMn ii,wb^f gD4xry>k5')rc4I$gucJ[f1K#*mBV}v9&UY񄥍 DۢRYH kBc [c8.$" ˀK+4VQ9z{ K|4q~ T3Nq ފ.}6@ ̹siMks䷰Rs!}Хh̓!8k=t^K;,;o6jόZdS{om9f L ?PE<scU|hR"ܩU#X18L Ek2epV\WƇٲ"M_E;~ [pesm:[[n}El+L+n,ҷYႢ~} v}~)οZȵM Gߍ\B.rxc鎞_fOojg6>gonZovk͍iVhy롐Wno۰1o#ۮP .]ΪZ/{| ZIKs1ţiZOCwt[?\+xA׌[=z]ϭCq՗=7vVkq-H&JK>s,rg+;)Bg??N0U R~̐`Cw^F:`7 [RdnvFzO稕'017Ulx_}^Ċ>^ޛ.lx3pF۽v_0CwX⁙CwW2&чaMJ4*v`9\VZ7odH/D xJNf/;m8,YYǔIB:#h:vD:Vy Eґ"MMߘnŷں&^x0ݨF8SK7GJ@+4M 8 0ͬBNՑ󝻒Rr \J_U W/~\q5VDn"r+]-'>\\_mfېցV_7^Mf,.>VT2 N(:<$`I[[rˎ5, M;̠28`ePSp2ȬZI2LY3h6nYc0{I]9Zu1JsFOy[j3Fв)v|4y )F:t }OnC0N%)sgb'eA ]TU2xC6ak'c4[$d utI){E))K6AE4Fĺ!>{;I;8[ȽW:Odx!fm4Mu(B$6sQ;ksGz8H Qx'JWc.@Cdc [e b!V^D I*e䲈d7%יNP$t`\f2r&ΖvcB7Jh>?ͅ?~>4Cm=lRD}_/N둣ngic, QK% Tҿe`1iB6&ڄx n*irV' uUdBsr`iUA GJsSjYY,x^~ue)fuVͳ"L>W#&6`G*NIeB$P2J̆IL3"͔ӢjlV S|W*YM.9i@!VP"pQ(%h7a}WZK͊N4[ 2\gY':C$\)uWqW|zx5}dVyD@MK:gЊK#hY#ΥKpIDegژuQTM&=Chp2=)dI,E̙0htTmXM-*U*laH[(-{|"ט~,xC-&_&a~[b+Qi#dP)"'TZȚ{.y],hP .'@%MCq\Lx)k dWm&iu)`vjE-q6[l?K.XjX64LpVrOc=)D$n!] ' }EUa&sI2!CE ٚXԍ'tYEȩNվ &f{; b"V[E"669bJɨ0RcYZӺF+<dkZD򍓒l[4bg|eI C Ox|}]8[%z՞dA$wLgVclE.[<9oid)xmU,/,-qR{L94ZBDbvU]cVclLXEq X'E}Zf&2Jzxvn5ݍf$7V^g7s4]Rz>aE,M_`ܾĶWNyOWK'~-BBzM :N(;v:s+ز`uv̌~;^ ]~ `?A|}'dA$r rb()] 90"X Z徏,~S:pGF-ܯ8?08.[>z/.pTv~{їrx:yHR"hM#fPcNYJ@J:Ƭ3)4)K""<_04p7<)8_SK??}8]?U]?d'#GclՇ}Rj>٪ZT.p.8}.˷;[u$3m׿nMj[“dVU%ѥ2~_m%YniVY;>ªКKHwem$I !,13^< B4)ȶՍU$TQ"D2""#"#ةc ]Q|㱦:>\Mk̞$ǫװ."nJ-[`9`RN(~hb 8̤ و"5dzUJ&}VN n7R'$D&.zĽj&o8:'*꟨$П ՔoL* LPƹj3z]em_ K)f0h=\cMK3|SK3lEKe6T 0T!GYZevPm1j;1E{v IQ7_W Ձi7$4ʊ55rqaؠ0=nJ5LR0-^Aͨ콡y6 o3cRU&HѾ[4,ؙwڍ/gmfmzuY`-34R R3\aCP"Y%` N*,ڹ,ѓ~-e;(n~˜K+$ bW uYpkCĖ#K]v{;J/qe58څGN螣tD/LD-tDY+f܄ Wu}fIl/ GNl.?8#Cޜ}Oޙ$ǥą/0'v2 O"bO7v 4*c߄1LH7y|2YșA#G]g CL{ఱ{L6C15Vmo&϶\J8.&"lӬ 7]`5x2Tw ?}x_ԊpҴp l,DgApcq7opxZ= *I +ofJ1W~ַ$pLp:!'K%+B#B: b2ψBZݲXC1JM!>IߘPd v|5RLyիI;AwҀՍ%urbˉst*oE[ow ЈLМ~X_-]?Pn(;@a"- \{0Q)d(r^#g1YQ^:O7>he2˿GM6o~;| -8&^WZTf̈qM3NS b8ma'x{ Z3X*(mxβ/* `̫vM$upw7`9b#3%#V&XmX&'X3䁬0AQ> Q)%ޔ:_@ܢwt;Nt!Q>;v~【`%cS#P \4w()3T4ĂaR&şdkKjLt9U|-34`ۄQ޿dmھm)+͏Q)+"0Ff=3%b#/3F)T`J R͹<ڍ&H pMAvGEgO>z70 B"I9cXXǵ,YPCF^,G`Qw1;ԫfJ4N'ϔ+z)k-|$̶IAcĂ&)"ޘytbHm+-WU(`>Y7p()&zSNq ,x.Zeb5sf"(5ckl֌*8cO]pp.o6$xEweMqVO[U5y|_jklB2(%Ijf%^I"Nkŗ ic]UM,g DY: >0G:Hpklև٩b%x,5W#5"4b!C湊(F%!4DZ!lQ$(#fZa 8pFV󒑘)DXҠO:y|2Gc-k٬xb5ų>Xgk\^d-EN/x N!fx}BsX^Djp5XDDhjsbEC޲>d M ۏ_W *a#\%lElfжÒR gJ*V(/SxhDX\qBր%T35.Zo/ПE] Z6׵KTEQ' ftg1k㛰h'/nRα @y ڜ/A]vcgoK̒Ӭ3Z7cZ/ڱPUmB6%彉gkr;YQĎv1~Ոl{Zw@q::o! t"0PJ.E$!%UT]_`Y=a.==h?\#O\"P KΨN*\ta괁"ӚpChC2u 5'}plk* P8}QΤ0NN#F#$^nظS bxŀ$ʀ( E=V*I))׎[e-N+Ry#8`8HDw#k-ulksi*PC{K-FkJ8^^Ho{l!eK׿λVht6.NLn^I ڊU jZf6nkЗΰ*&"TEjcMsP|j=u-_úԺ1uZx8#TI9uy-g-h58̤ وz!uT &DUJL&}K E/5ܾoOFI6L\{9L2@ptNT?Q7I?A8)!CߘT 1=sI8WKTԘ~W+ a) )ٱ٦BNZ%ͿT2EVІ,p\-lx;j}8EpkZW9r ^UήrZ&AmjAf3x]=7$w7}.ø5q(׵vvj+[l}p}>[L$ۋ|([g=N ZQSDgϛƅO#Pu٤oҗ8ӔT&bXY&wz0lPmsytk%&JؠffP4< 턷1fU#shܭYhmS;Fⳉ6]}߀{sSo 6pqS{)TT!(VK,P0tYVGIDϖZq@ xPḁg1+:PA8F{5!bˑ.82{ǂSozv£U' *tQJB:,ȐhhWWi3bԓ]g7y/=Y *wVys>ygO[n GwO#ɃabcQFmVp8jOR|Ȕ ϫ)$g^"g揌jpvQc:j}$C]-hɚ{LB`և'|zgjz aʺaz6{&E=4 ME\U>/HŤ"*"t33SބYgѺpcq7oGS',i.۾R201`̕_%-`S*dR JHh0zѳC=3!@z֥]l"P̄?RCFO7f?8b<7^ vz?5y{ƻP,t'DQݸ[QvbzL8@oc/xzTPڇb>V趝wGGѣOg➍}~ީK}_n}ƥ]?PF94ו1A}08vTQ͍Q:vF6qMr|txwinݿg}ۅ΅y)pږ 3})r\Ydڕ_c  bR%cٻ6ndWRC~qU{Uɦ*Ͼlօcj(meki.9i2g4F׍FwP&˵ARL:fEӸRգ,tY5aKBE 3a;9 9I6a[ C!fJy Y@SJx->cm95r@:Xy#C)qD)x -?3dJ!Z8t햁Qr3#CGWe)AX|c̱|2# $s-9˝R&t={$u({/—Ǐ㵇$@N!y04Єt9ptc^_ͷ+Gclm#|HiR)5^~&s uK_ͻC1\Hoogs_ѯ}5&[$9C-W K-8Uv ?éSVfF90EN93XZeP{({jG־Iclq)&: rTDTzʹXH@Ѕ> F]u(P&x9ɺՑaYP-J)~)$ ˴.0C:iOtǒ+gr,mVy&裰`x83ORi>ʫt▱ .[@C<y,![nu '1beP:IuVt .&\(wp)BSϘ1b"ij$s1b帯TWFW:-/՟tۦ<_y~,N-;h=XBe<ݕsg0jõ,H}HJg$7K$cсC] !~%ݕi8?RReqc*h%{gqoR0#4'(xAc+xfZjOE*8ޖBt\^]sX=TPuth(dS17QVH+1(_SRRo)4fٲʵHVw,]xR;f4LeA?or 9Ŕ_,Js"эghQYgݎEjYk4Wd)4nH^Iv&gTrV@6~lL/ d!E@G7U}9;y )1>䢳T|TrKKlW=^a 8h+'S*L]̀Z WEvTFH4OE^J>zoL?,lӽv==Dm4h)9tMq'4F;WW @u(F qJ Km{G׻KL?GwP_[⮧V>dwyުq(>*c0`4#k7NTN31̱[ 2Fw!ap)1(LI8FҎY | F"G9 wbnyrP0Uu͵Y]Yiw5=+6S,+-[P&}NI4T:,1sƵgD(Rnor0ǝuHJ=:_u`qMb>O*#R"L' BaQiƝV``#AyGACcPUnsmV{6aHl0c +Ar=# 덥{D^? שRMM6[11PP b ԇS YB$hn#U5\#7/z%q\P0< _8\-D?K'Z#e($ɾk܇R ‹3Wr=)]GWɻ"O!<0d֌t63NK tp U`H.0gpmwu/ t>}6[  ΊN95fr9#lN \ePT^d,ίCLV)pݘRLg fٿsg|8~P=-rqNZ0(|]q*="1[S0 :eI΢bf`(O#@R``X)Eu֏'mhD\ ̍hs&Ozc4 x(fÍ GT9yG;2y:FJO1fDc YȌ!VdÛ97!T;PGU 5xd-pڭs== ^t ހXgHIN609ډ\r2յ{uD%]uo6ǘ.0cOOCP lp?Dw|\pCսʭ,T Fj[ [r ,bw0q㐷$y8*_<-2T J-I2T`#x0p/ؙ&Fd D!e֠^,tk3@ f 2b hFs+%e?Ck4M9}2x~nnxs;oa٪{^yޠ9Ϊ}7<^NO쎦ZyÛI٦]Je>gJ8(/'ISԸdo>$ya:FcY6of>n{-/^Ԃlқa4niot bMxLfK 3ꎆ{QIFkΚ^mzVS~ZdhSu8:wawx\:3w)eZ,ʢo44mrP~6WH6<%& ?}gwl&GR0)멑Gs?z(y|t &'Dä\\ϧ |RJIRFqs-5!'D8c$%;,8SGTKKΘ:@xJH"a" Hm12y%eZn^CQ{0rHRSkE:c}J!LMSxbkƋ7Ag7dX\(hd3-,(2J FV`! NDq(e[|l%4eʦ^~&u_ڼ+>EMDaY#m} F/ݜ[n;Jp}e!AZ#! Jku2\J|칩tT% +bW\z2*Q+бDRN\};JaB*Ɉ+ Wv* KuT\),ŕ&HSJT FΥۢJ䪓IT ¯cWJ%*&ŕVjɷ DJyâ50\p464< YiԮ,n/St~c&Wy9bygt* ےik CRZfw"{reqr6.oR^9%^sS<`s-:2|Le a1B^\qXtKܤyN?`^_+X"aƁ$SJQ`~}T.e)& |gS|/0IQ?t6I|`^Ίxtۋ8!x:OO"eX_ו\--~~.fgK\X¬q!t43oevl|^MWRA܋7z ϼ G: v zQ.t9'RR-xx݄!=@/dnRz!^fCrg^Ml{ȡ},PաjC鿣oJ#ҿZDR&(@߯A/FC?vfb3Mz*3 |Ka`UUT&xgU F\a߮޵5+9B[jwlU'/ٸp$)=p(Q7%lY1 Lnu=|ѮޔLj%oz3|FHyÑ\eEnNEu|ň4a]֩ۢ=ioSuD<_Uns>>L3 j}>-I>E; ,{ظhK=hx,}e<8:lOq(G?~nBLW*ܰTv`85Fh24_?o4df A807[l&O~tW>ɚă14ZG| 0*f F8@˽;{MtZ"zQڄ< iA$arhӁĒbI ԴJ#Dims蛅;-H#k>Vff/ϠlH106w=!?)3il. XdES=il =/{͗p1{5jh2.S4ѢcwB\%: GBJx3V칋'.W{J>~RT\=3lΰ; w@|7 w@|7 2 \@|7 w@|7ߝ ܓe@n ⻁n :D6EmCQۡPv(j;\z.F hyshz |F\cC8̍uL1z+NAn{fY}+8z&\2GbHm\di(d5PnvҞY.}@:%JAPЭӳ:gv-Y+"-OϹuFVwz;z& twم:4{NҽwteC(h.bCkPH6s7Gܰ9wsI͍|r.^3CdcBך[e b%"j-IR,S>&E4HR ;ɓC#Ӂsʞc5qvLy ^gŗNn3[:f讯GK^![Mݬ`_YZ VGҎէcV,a 5ܥGkBIB-VصoXcJSu`Oyz_`Qo2O0%ȝJKg +xJiA6We^qvy.B-pfw-p`ߧMzABZǁGpa.84}íΎ8q8|J;}н!vyk!Ш9}KVal#ӟ7Nn+V3t0D#CQMeXrtX-iVKujRaA?Ҥ#uQH– Ag G]$KȊG)E/AL3њ "Gp;+QW5qv<'t 5sǗsnu1@C 98irL$}JF0 mZtFdrRJ Vf0Ȓ˙4Y@!~VP~$#1\((A{1 C|Zl7t>kP'ׅW_]K4}l~Ieu|O'wH:ʄIX=4ue@M'}PR?q/N:yՒyӷz>SI̗5?M.:T:gЊK#HGRGDE3-:jpF hu("F}򞕚h1f' L$yH2(2Tmd&|jXXM3B u,<HTvu]<"->lSuSοC~ wy9Z KrV F2 垼40RD4Nf5\X6.妜 d]Y4`E΄]-pM؎Zpi:Y0X8Oq%<Ԯ0nx`Ӥ r4z2|y*5G~ qhJRÐn7yn}o-vPk>mlKx_eOLS*2Cud?,R880YqϒP̮>zzg@W F& @G9db)[B%ْ TLN"Oq],z~<gFo@PYY"Ƌ BFM$߃t:SPiiL`B!5fjV>}u-zlvWPmBՎ.va=/b]]|>׾_"B%:*%J Yd [d L䥠T|a TdgipBIJ.T=IrU >vϤpWb'i6?.\MfZvSͨWW$S(&.i.he*i7>~L%?}w]ο\]𱤀mJ_<|_\}'w7SZ<tC}O.KYLyIи3T0=C X(BgEXFB,PW'EuGϓ/9>vz}3.ϰfgR0˽GPxĿ^ג{Avu ?|3nҩ%$`u-*k;J{Ki9{pTo 1D 8N*_'s_`@\} ̇\zr)vvnIo5ц3w} z]Wi"mFo|_\_68UTOQni̷tohT~KP/.fzƩ10[k)Ci@z!(m#Y0 j RWH/Czi^A3v$dO!dѫTlJBϴ d8CaO9FHpWO(Wpq[w}={1Lޒ<'!(:ȕ&fzFODl[ n:w^tuqq>S9X=vfwI1_BjP`"oJƆ<;l\JɉVB|T>)CUqrx7zS 2^$g㋟Bƹ8.8}.>1۝Ś8Ubkq(U+MN.1.Z-\OyäpG%E0H5.Aع*&va+A+o5LjrgwMkVێuOr/IQ03?oG̈H/8~#$,Ja 3ۇpz趴2+o fd ܄o0؄cE?p;fXQ gZ{㸱_ii˗cؙu0I& XC /IKԭ#v1}[-KjKGJ.//>.ޯ[Wwrw#a1< 4150qQJh>䚋d')GmbtiIiCx ɊqO Xs2"YH'R'M& -<-%g\W"nmNx :JEWos^<0,Wϑ|WM=*,IY@g 'Fg 2s$h|jZ@]\]|Yͬ-9+꼤HpNJ(P6ٻ6 >;-c௳r_Y1+M6,;(@ݪ\hA5VfQ-]v"gXBh_O/f>Wjڪ\U&' Sl_XgTF6쓌!DCHRBI)H:ҽ^5W^(c?}qK}2?f\ߑІ*2?9IyQHɩ7kjޖ}wdZuKߵfbPms02'7\tsnolD,*ML@kjXuVCOOQ:9=C<)an[Q"*  QR,gї:qr{ٸI웕v~VJ'Nf}7].,>[k.>=;:~V};ShrM]3Fii('ϑ+BZirȽ\Ix^?_o^u_]jSӟe똭։mt`\q]gw[𽡅Ϡ,{mx|~-$GI] Bq Q( k<%Tv^ 1G[7SjB{Ώ#}vd8p\RRV 42;{[)F+XFmJLƓEq=RH?nQ Y4l|geyc~̽~;CCNћɓ_ O}!h“gOTgO~' U~,G󓋇 '9xӻ[xkVO+r3ä4r`9_tVk+A_j~%+6=ncΓ~ԓo{xA H6yѫ{ZYA|ʭ$}5Wڝ,v~Y'w>=-Xq5~dv]40y|bhEkӨUQ֧;Xl6H_`eU~)i4JCfRvR|Vnl]Ney௛6?:zvbm~g>Ħ .jXJӳ1`k|w3]sHXDI[ho|39}wjdt XGt߾쉩rZ$)dRVeTbRQGyQ#W#%jӔbsZfVD#Pqd? EMI< ДWє(d?A~?d6 {, ̶׷ό{Kp:!Q˗z)F,(e eթ8R>L.O}whim$]p>vyXfBCe69s rOw& %wq{}rhmn6?moZ}Кm>U5W+Z\j`2xCrT8I&okZxj]? ](i!X>;=y#Sl?ke痫. qyEW9GNl~vI&%jHHqd@jVܰF_kx]Hݮ}k^Oq /\MOL0HmUi k+f2/@ /]1WL?: ~Vs>ll?_X59;so.1zO?%IDEzZ+k*IY-GiYFM{Z^{nW;ҰkHO.v`cUO-~{Mf|XKѪ $D䈊Anr"ht'-%h&(A2]/s}N.I_!9 ZL)bJYq֪"+YRu"m6 0t83*e%щU9ёbQ|begg0f >^=K*y2;I !S}:Ф4IeXPQɔ* F"8}₁WsUEj'~%|<2`aDCaA(+gLJJw>TlZV:` Q2hOۢmEHqL.kqsPټ%U*G:c$(H%ehZdJH<@7J?f< Mvc%H ۉ2Mї`8Zį#t<_G H&iQ.1cFaOJ=j^dͬ=좗 U"\ J{+kRscԈ%gaE<(pURȪBT+* kEW KN"oDBR/q h@c K.!"$ԗTd2jn57#2 nEf6ۄs"YҎ~^[QH9m^V`$aª`|.,JwyJ6 \{#B|t)]L Ajx1* 7D@r!jT@ @֙ R{0Y$[.F0ZN ;f]xl;$,53`0bduJWp|Ӄ &e#s0qصFtF 39,Di" < =;#"ʙa%oCQaIVc9"agDH! c5Zsf6?-6C:f0M `fՀ RTiKo:<E8] 7*-Y !):кf-KP`- \i BhӰ㠭zy%7=Ch>=~iLjU̵D dnT\L5JG+40zKp0nHQ(8Y:zm YT#yHY5:`4h;Gp31iFe0Bgãa·fUҷ9p78%<%2` rhcH|#FlosN,DZfd*DE2,m􊀔kz -D M?_!+Bri%s\ \D :3U {P(hc)2Q, g=gU0?y ڈ81ul[EH&AH@pQPA(R䚀vYH `2p?e;\Kï㒭gtfg5^gM'[`0`&o? tAkȟ \ 뿳Z0ۮAْ %l>xSI)2\vwZA !pc) aR$H-4C\vOhJ07LJ,:cMs[mnQ*ݴCBru0/"̎dlF821D)h=)Kh&, DHUYe}"@z'"BpjJŖR 7RZQ?x)D#.Cx>,ZZC%g 9_+׉zl4 taN$$o0-k 9RzO_̃JzA`'.ͅ`})Z!ϝR[W 3VFksrQff! Ñk`NFyan iT N(HFth~f'x{нWvx7r|_rk-G1ו̘TDN+!Kʨ564ܔ٢ !KYsv4_F3.:Žzl;Z"g7Bp.nkI]B3N[r<ћ'A4A͹T MhF4&h4A MhF4&h4A MhF4&h4A MhF4&h4A MhF4&h4A MhF4uMܮ}&h^ Z+ٛZ +yG3ANü]:Wd~UAAdtoUtEyF2KG }џcގuM7Kvg-r6GR7]YMIJX%2RY(Ӑ;kDI;q`5lGIk9#_WY:Y{\6^ܶM}TRmTs]|ѥ$z}'- 76@)n5)MH g%1bndLyorc7:Bޮc:o~:&7oSt}r-kv>%:~>cnne[W/c!:T0q^3{Qka{^)-^іy{ŝ^.ORuU=HV+"A"D!D B"A"D!D B"A"D!D B"A"D!D B"A"D!D B"A"D!D_A$!^_Hj R\ Z#"+"Tb ZDn6qsv;!&B͓ NGӱ/ ٓgpП[f[v=nv6 ^Bt?q={ j_S~[2ksjrL!SƲ5E=cx~$S1 G5BZSHUt8x E.Ⱦt૛~d  ;\k*võI"DJ~$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ IH:@#h;%t^+ŐIJtI )nN:c/G,A 4"9R? pY\: 6S?>/Eum5-1:% B(Z@6aACEJ%/%*`ړ_ t:5u(Foo֜N<6 la<3D\J`'ժkn7M[qv '66 &{S/fO=~u焻{KȭM}֑_gkhKI7MGo?=tInٝ$p$YO:OHV8ʑd_yǯ׌kƞ>j~3xŹs\=OWiwsv/XD Xܐq"Jt )H@JR:ҁt )H@JR:ҁt )H@JR:ҁt )H@JR:ҁt )H@JR:ҁt )H@JR:ui; __Zj}_)_Կ~;朦@C]^T@Eœ=TֺY|E5 0CE6y7[nT%2{A@F{h );9rt vȇ8}e>|ɉ)#a?&;9I*h,񻒄lhQސѬ9!ɧjы^بb;}`0^E]MDh"BD&"4MDh"BD&"4d9TРǏrS?OonJA^$L8}mmݏٲ3l9֞QfkeWuJexTcO \{&#b61XMV )7M[&͚;^A`(fx%mCqzwo^~m]'_m`ؕk/갲 -@5_=O F~)~A?0{`4=ɝuۗ_g[.Ƈ|q]]:e!9]_]nr/~5VV4nAVvͺRҕ?`&,7KOp{_+m^hת_ONd+4v@.EVf>N~7T2(cD$Hj )V|(0 ϰ0 {[*ڢ+Q☕uGa#T4LsmM?7|Ԭ9 B}NXK[[Jr>=Iºn=WJILm1o/CH zd![K3[OBoV'D"˔ .P@2pY)Z,p*URi dmVNܜe@$]ΙQ/RultDG`jɳ N L'qPn;9:::2~ўJ ֳHz+L(TQh.I+b6Ŗ<#\Ź3]h^CK"B b*d}* <temrln5gInYO43dh.v2[Mwh]Ǭ8?Rޖq\:C,1 kHSRխK0|s`ĨxєٶV{pBjQg} J_k;pm}g~sN17Y(YuAk@m6@m{tW~@>;v_+*%/"R Оf|Rk’vNya阬h5S*ME0'Ǻ IhXNuw--K3!heY3#2iB)#5Ysl.W .NZg^rh\qq[ =?F*+hhA80m.I%͍с"/i".KM6LՏv7nt4(zZX"QgI$Obх; -M_葞nof!/3-HA"O:+<1yhdLݢ]Czs,R& 1GbM09w9uX]AIL?>̖m;?:^Ϸ3~7o`E{gHOxQ50:/xƅC*TĔ[n\/wg<% =#ϧ˅ FX # (Vf]ASTn u߾k WđG1,\GDъ &>] jhpTB,i18ꛍ3)Dw a:2wXKp%~.BzM/]W:6˄M 5Pb>I$5#*6v++B\KkP v2m^ /㤞+F9| S$Hj:wiqé΂;ρO(5qǕ(I]w\bՕcjNVT ?s0mLxm~vjֺFI;{Ps/ȯwzw.Tu.1,5Ӕ``zpf|>oڡX12ILR}{wraRw7٦A5&UZYKS8 0GM,GX5{Ψjgoo$[Awqj0 ڝ D. '] '`kQ9WTJp`͕>w1kGƣuXoAJպVu*—[iox4kV!Oȱ>ӌ8%f8-Rh 0B{@3 .#Q0!S&aJ;f2Rʃ$ݤN/9KNS>ߠSDIÞܪ뻳*571ͧ iIؔeF3wZR)ݜN9it<$Aa+E`X30qg#2& 4>`znuMNn-N3E*#R"%1dNF03,*q+!ʓH&>+%C㞆"QTBl"AO-˰sg$p&`F1usEK="aElO햦ZBB[  G= V b n% f ]AZB[s8wp-8iݐw1,?qk@!IgP[N(}Haa=0ky*TL@^=MA hiCqӵ J_Х`uswѕu3?F6+t*J~N#*wKXh 6FH[kq\|fKTދ1(NYiRlj1z1?i:6I 3A6wHsIM^5n^P_Sg4".hF9'1 ~\BTfM3!zr~;#&n8H Vz1Ĥ)kb !siI[{( fw!F3+3`.8V\9pkգFR7F\]t[=M`ʈj0m L$nl!K{ɱ6$Qs-9˝RZMAilfsEZB,ѝS Std8lJa^%0& ^ wE% @UޡMV<>xYrgm e_B\Z ne_ z7M}57EK`$ox}Ct[FgS1Ūj0nYF".;L*ijQ sӅq3u ~ւaN(Esfͭ$G~Zezg@=b3K+ܯЃܳ4ƆGbB[ GE$yNL Auᄀh &9&&z EέRȰ,^kZ )$mLȹ{z95F'hf?mf,%t"3}QUk1}`9aT1U~w|)r-c1S.N[x LqY̙C:<NcʠvNjdXjK۰dL(wp)BSϘ? 1b"iEٔﶘq` yMzR@[_ ݯ>{)\np-L24cjjyN`TD$a`} 7(JWkeWrC`yW0"G{?Ófia(REop0[;?@>Iǃ0r|c^j'è$W< 9EIH)CV KIuj{giV6U>eaIMB`sjm[ tX^Oxw_Ii:9zQfU凓ʝ40۬h<(=?*qu6RӬpcv7z;Hm_OVB6b^O.Ź3?,BHn9NR7]@QDIUF{jXiii$FtYVOk]Wf.RNq,BsN;[B7zf!% 3LCY }zӅ\yg=w*v([{pdqpS޳Fν~k"":,ָ׉\ɷŽjkMv3;` G`Kԕ`pQW\AE]%j WWJv+zɺ[lkDBJN]@uF|Z`LG]%ru"6]]%*9ީgSbJ'WW!'WW#WUW'rG%ٰ+~uwjX+qJ4 yDxGvWMH):ˮ'˾>?fQY:@;X&j~WVSK%[JWL>Y! G"6AߌJ ]vkUOpFGnb ,CgY@DRrSv* eǿO\gG%*iBoeõ^A'@])}> dd9PL!/׷GDNR鷾w_fdj?{Wi9F'G}36/kD-Imoq2J̖21HGmTH!TraTK2ZRM5,49a \gfg>D SjHNXVi"zrQUhvar9R!։VͥBD+uVZ+z kmdlh o{_>R:T!1p"IԂ)ЕrRWlx:zTuN-TNV G+3$ -ƈ魱BlݐUzonxb&< VDnJ2-*Q+զ+RcSW/P]1,[lr9ۢtUשׂ8N[r{D HlJTbSW/P] B`H]%5u 5{WZg^J)v+`=i=,֨D*QnX+I8]HuBjmJʍ* ;uՕ&B +C9#0. Zg*٫D%jHE>+D:*!q `K5:h 03f@ʎ73Oa^x)rE.?{F<%3k4:U[* WI1Ғi %"Q4g֦9C f׍RT1_HIFUÀj4cÀ2Ge/fP_"a~돇7qS4lR&c2a P )c6)x)a|GKe\p㮗;)Q=o^^Qs]-ZGE2eeNY 4bU h %XHdb@gci (Kkk3N)r&"?P#jHq f9\7>\jvm:2&~ usׇt8\w%V [nu|h6%IkIN3 HD0`L9QD(TW QiZ9")Pr'OA)MbwX@uP J֕dҔh(INȀ KY=c0,IMma+:oѫIFc,˗+sߵ-#tg ]lR :+묠Y@] Gus~ė.S;^cI;7LZxU`RC:6K'^%Eh`U_twߜ} шy*x(*d?\2Jc!6uԡMbnۍ+>-F.zcle1-kFE+Uj;T6=”J:#0 ȪKk2MVO}F}|i݈(ڭeTIOt[*F W50ϩ\HF#S1B )O/&>*sE`XA0AQjXd}*S1BwzD:O]um-wwI]v5{/e^ };X'lI1)I%6Q'Zl>%F{"!(M6߫*>8i⴮W3hOBQB TNY$JjTDK>/`Vڒk2d *(-ꂹ(/*YPke3$L}1n6'n/o4"~=@׳Z?fa;]dyYf<6[8}o=Em}CV0Ȥ .@ljG2Tr16GOr6)[h)ldPRFj&M,ebJhs5OԘ5%ԬJ%]@cNQ|lk60) `D}4>Z _aU]ᶏcgG{nje̿t5trq;u_L_5h][6vV\s?ϫB5 V_?e,?o$p֗g7?Oa 2r0?h8[}R3E 7}kNe>C7W|W=' P%?mGā)`釂†vņݎ&b@Xe~/K Ħ­ =|?|Y/q{PwWjky;%wNP>^_ Zbl_W\z3W,{{ƫي]{~~m}Zz_oN|_O ݪ2my:qlsnlvdwdm'Bƞl1xAg|5kw3=(;"7`&Oi8̫s,]x gFٹq9ݝzqe@p≧]+x@Y4Fr%&쇨fs>\ٲ֥aIbL^c4Ó[O>_xV'M[e+P[2H9%k<@ 9S Ad"{OCr*k4_,:H}Als#'LceDU (E 鲄R S hB"ۀ^XλUBi?ۭJ6œڛr~k<ǁ놼pXsJ5-?iʊ~h %_-+zp+v3?ϝElccuj)mX-+CmGb0xLdJkw82zb'X>j.3S*kLNc@e"IXDZxDIOyy|wӷSk,YFT)F $LR;eK!fI AYl-L&Uu %XeMU٥h;u2ٮlWs:UI'޾& ?oZms4~,Ț g46Ss_6mL]/%k;e- ̤+}dž(u>ҕqrȱN]|+Ϛjj*=˺]+9 cnw|wN7Nrx%,$&'Q8 S TQER 9ގ$\G.A<`Vd^mv򸼑E ad o~i k i}~SZ4E^g9O!>| j$9Y4/i*vem^|gãǔ.D"Y4Ls ZLEDt+dx:O',k@S&d?`z+0lү/o3;V6qbGoĎTFw>?#= Zt|SEwbbB%sfjp858k\;.XDH*!3WL'f b[kO)A:,QVLy$KRojٚn{B/=з)\Yߟ^/PEou $Z7T:٠]bQRC,jpt Mޙjbgѓwu#oD%{lhg*Ql:5x+ Ma{5'gnSJ[e6%TCCQ;VnVUˮ]Rxm6 U.:* "OAf<QsioP֌%3(I!lK&F+Jdbꗏږ.s *CpATI sEyZifyATGXb)ϋ-_R?dіRCRR:DoUF蠜dj]љrlg:)aӔ5u!ş5O){Y``%RL)X+(3l5ÊVҼU:._jVKZ2ŷ]P2|}렲LyC˭C,ُ׹YE{G&F%*?ţfXYK%ZbM2yKAU/0eTSd4d ӧiϋ@2HM<XNTs.l4 M02*ͰdmnB>mϕA_iȻ]~[ad_ry#(jN־`Q% #C貈E2\16ElE ?j$ |YYaSk0LVd vBLh1b7aWabIXP&ԞޱIyÒ,`2f(#APT5ϼu8DuvSthS<,{>Ҁl!33d+:lkRU7%_N"Qo'l:75a*0DZ(#qBĽqn«r%:x!pZKah<2b5A7!׶# YW _-YLH!f X)F'F"6Ff9_/2\wyՄ.N|p,1 >1("jO$FnjH:*EcyǂͤX<ԍ@؁evly.s]'MяoZ} ˫Y^|wb]w>ѐdl],FẠuf!_R}NU4غYRYKPg(GcC$K%R)D\ ,ϯ[6ueEk̟>Ra/t|u~vյ+ )d짳cᚿկ?ͯw:XW5JjפWCP윯ϟiݽm-y ǶY9 +Dފ99Aqry7wE2(0+ʢC%dH(~uJN]uJRL"mLy/,ɛ/l6:gaeYjR"٬bU*$F6SF;jXi@2 ĕsa%2U-:Ve 8pV[grgtae)a䝠m_i[謐eAqv͍g8z/ļɭݏL<8gx6ETf畘4*֌״JҪ%z)#4K.!F Ӕ:oyd*wp#TCi`LND<sX>G6v=A~$zq' SkcYwPNlO=={u鋳?ū3L/'g_P`mՄ‡؄7@o 4ASV+W6&p$p k{k@~JM: 8N]<8 \Om`~Yﻛ'X1"fXپ+Db=E/*+BRiQMHM:[Ta FqvN0Okexc?k:{@%Lx<YCJQn&#_/ 6Fzc8N1ڃ@ l!s*ޑG9-XmnyTa}gQ#1kН)C0];h f!TǓm mmMUAX}yݣjU7^6xpi[v]s>wm>A]ꨕl T$G6Pk'9YI pV)z&} ]|KQ`QE}nXvoۺ˹mtbҭtR=JEhRO"1J*IU\5"[Ib1{x䧓£w?]mpyr{|W1Е7/7VŶV6#ff_ D%":S?qpT*S?W <(Txv dTmZ5]Tۖ[ ߒS\nQȆJu/_{\Rv~TW^'~Ax2=7IYrRAL^LO^vX)nv.p mCLu`X@J)HICҲ{5tv_BToZCZ:4y9;cstt/8gN&bQ;mRзo{K,5d,LOӯd-(A>pF=RJ74(1H?+A8#)AT?8 2LH)AH!3 Є@a\m76,Pa],Î "k'%%Dp) s来$m[ j vRn"䫩ïBW Roțq|?^{sV By3-%Q2fvšan# zϼ0o};hۗ qs/ ˪YbΤno7*~2a?ՆoT4EiނyQtD-qD ZҞ ;3NdHfA M_*2e6 .0!C S(S҄UP^@R f@31Mi0LȌFsPm3"Z[Cg`ߵ]^EDQg m]x۫4Cr)jyLZhO/MTw{l:^Nu1:-[/9tSl95e -:w7z5 =ϵ\FCfc^e̓EZ2E;ezyjB̍zItiï`m1//OR<+1N,rY%*Qr27bpr<0NG&@|d4ML)&YG&F&lVZOb2d>-$7 MW0/t2Áo&ӈ>Wф)as{ յh, 80" *L$1U^z$ܽ*?W}'խjieo^TqYBgY^#3K3(f)3Zڸ1%ɍfbهzW ]!T*K4aAw&&+DlRh~d ̈qѥzWUĖoXJW?bL+3*wF\zvqdd/~@qřDڝl ]3"o,^\8JrlUDUDww|W[ͷ]\Q\INvK\E9q}TF*b+]?R K ,qbWUĖn1~Hqy,^C twN#ʭ?X1+`̌bu~8(*{biHd<uX+jTZF%Ek먕AĿutq!$1P*<7$ L2ZϫF[γo7/#U@N9X}&^ݘ 6(ObO"o.P(fv_TPdHCBn 9CQSbTi!ߊ=P:aQC_8.(Jzs$vQIw)hU;ɫRrYԏ?^!1s̶kX-Nm)lA<>K\CX?{ߊÒoe)ɽcAF;$"L쌸r+*bnURʽŕmm3NJڡ0I9muǰs:(`boF*4[RZ{AL(׭'\,zxk XRE,UF!>djPPXFeH2Rdz͡s%*aJɫoz>fSs:Q+NIJW Xs1iĻÿ g~A$Bx33sl!4??><9r¥Xz CSr͘I7)cSD0%KAcU'w|Oջ_%7 VV"3L0ߎ.geztY>ʋ2X|R8*jQi泿J.M!H**eߪ Ittt8_?qSx(ǽcWj o焪B/+.` )_ժTr,jcnSp}(߼.^/'-0EgtoAu/JR=휣$ ɅLcP$ !f0b9Ft~0,dP0R1A&t>wp#TCi`eDw5ٓ}#0pTx'9ȯ4a^ SkcYwPNlO=={u鋳?ū3L/'g_P`-l &p]>l&4'Vx[^ * VG]|97هG^ Z0wo/e7ɰUzkή%fWh'X1"fXپ+".}wV 6RiQN*n7 Ci,:[Ta FqvN0ӝ`7 ~t0,J 3^y:HsJM7:G&x;^m2,Iqٻ6v*~'٫ݹ 2ƑIN`KeGi[v:8l) Zs@% 39`+Yqn]6'FaX 2$/ZM.0TeG"/ȷ2`iʷz~ʴn4PtG&!ԢD{ӧ3,q#OKڟHD +EA5UN1WȆ=GM V\;L)n?=fi\I~eD hWm~lsiv}47<0Opl˅떏Un̈́Zb2zΥtTlP6>|8jm1йh{>s=G6G?hD3XRiHL>ddz=-1_O#9F109UU6z>z.n1Ko2΁6ΥPCՂ]vYs)T9MV"`Tvyy@2ާ\ι?f5uL{8"͋}^ xcy> RtBC#ubR[z !&*d2 ޣlZ /":]+8[ mx<b:A-ÛA5gӵ|_˟6ր-bÝZ z~z}Qfۧ;Nwj›kZ{NMJo賈1k{y#Fƥ$e>{D1 H5vʠRh4T?Dq8}l9gey,7)[!(TNH uY2UR;,ꐹ0ϔru"Vґ1#LP K6)&dГ%V7)beho™Mܜe~e]3ET^O޾1; GA}δ#Nuzwr]A^{jhVJV+> jT$_8Ђ8 F(NmҾ3>&΁+K.ِ֠f4o#X\?-x:b߲[cqU>MؽIc =]VûLktu`!wh#pI  /׆փ￵ww/9[{4 -l=]KϷwJ:絖bGWw!s^',} ɂ*+"o !3ʥ *Y35>VQa nw)W//uWq#3P-1%>Y+,^Q%ޗP)ʪKN* A&M9:Id稫d3z #k-"(цuNާF|.\+St o{X- +*Q0U(Eׁ[%|^k-]J.sjG Xqi#cN\(S)X2˲Ǒ"]Dm0oMW ޾E̴jK@_HIcU1& VA:|HayWDNw'ͧc$턴^ iild&i>XpgHed 3p)b扟dJ51/JVp("Ě:s\@7ʷWy"QC.zc)= ڑ2/KouS^ ɴe >&Jcp:1}Qr:DBm [gѕ`@Os"y]bL9h 4j#(͜Ѧ ޔ5J̆c<[$9~ARx2 WYcJ.}~,/^ǁ5{ :}|*-HlWZ|b;\}pj  h2ʇ5Mh M# +"XN&'Pkcʬ 2qD:(6 9ɰ`"TP5W^Յ@V>+$*vF90n(~М.gVBzGf tz׫e/~WpBmTf!D4+- 6+.:B6oICNW8M}>F*2!E1! gqZ]DjV(T/qkFF\P\{5wS,o+&/CKy;(`1@&!GT)eqǿ;z,_8jC+/5 sZd]j՚U"*q_0Lm1;[A?tDEɝsXj):%|9r2o;sp?܇s`FɑYC6hLZ;miJ-'cUf(xόe:EbzJn58<"XaUU7*Fe&\hݰJ7[mf<Zk Wd !tڐDˋrP/~,jw/ t~~jv``b\IPkj]I,^>fr hT+М=bLlqtMrБCN6}ͥ^bk.XnX6LV{Bf<(d1I"$#HŊF[DĉÐ@ 3V~vUtrBV2Z?~`!Η5mڊ?.8qjNgX%(3WOJ~W8"Hk!*ExjOa2 `APGGÀr0Bd'L Zu@& pNDWV3C*Uh",ډriѻ!ѕ,PMVMcpD@>%)qGm!H>W󋏭ͽCZiv75~f߆|;7<5*9VAHFl\łSPJM.AN :7cs1:l5=yw z08l]R.5U)l)f5 ҙ@< S>oZ[ʖk}]u1)9$mF|`285HۨRT CIiQO>.آ}B=~a /0MW~B. \x3-d1gj"ݲ.9cRӺ=aNMs}O^G6oOV% ^KU}\OnHn׳e.onu*jb[cTU%vA3c .9w:oL{:>9NH89]u/9NLytWnZ7&rrv5&-../:  W3>PVZD.[ !`VͶ*bjOX˕ٝ/5*Q T69b,oƖBX`<4womЦC 9[Erp-9$%ѲlqBw6lrˑźFz^ev(7co(4_ju5(M]5&f u.6?o9q< ekr*]@?ôqZGO [_a5M4ӼfXkb5:jBjw8q:βoP BfC|Yf@~d3Mw߀,k0單[2_4}u)&m]8uL-3J7:w Qvmm:\6~y.*xT^H咯j T^ˇ5n[>iQMLatE |auC*!1э;JSܢl{dzKS"B%VHm0sZ@H45`R[IL}YLHE Uuʎ3h>m :'#|o4{][JfiSN#gPڔ'Ie02bv3Ks#x7%850.|)+V7|mvak.~8ԭ>ͪhAVEÊ0J qGX)h%S$6 2I( 8DYdRiVJK/X)W=ъXrKrb:aS2l((OM Q ɘJa:<<לּ^7Jn3.Q,KY"K)r1iΒ4N0 w pUYFhG3g;XS] k8bҪ s܏t[|Yҽ~K绞wJTLeG26ܤ! 3aYRUVlyO{l*o9^n}7ڻ>e&9`榢|U]^rN;/V^1걳&9Y gReL&<(ϼw_`QPA9<3OFB|obr> : d~1'Ŗ-$({>]R xrXo qEVRU JiOrVy#WkE6޺BJU{ '#B`uBZf.WH)d'W-+܌ݫ+6B\i|+7]NZ(W֠\!3 pfij\!%7\}5rev,z'pO.W;z`=Z)O#W(Uì+SCL'W=ԲGl.կie@x?{2rsS#C T $[]ݺAUgbɼhy9 &v8>Ԏvʧ`42"KM]H @t,h@ M(yLjPΊ*+o{^6ot z[^,!T57ļ~U4ȻCuKY.wQԲ+d/BqRQ<9VOqœCړM6zr\ tڴ'ٓ6\P+`ˍ7r"WHkiJHE䪅rSh>B\f!RtBJYWm+8%> &M!fjtJFI'W-+}-+ڟ+ڛvs<ҔNZ(WFEGr^WS\!huurBu3v"WH+L )꫑+c>Pb&+*C#M1s%š8}|y=ԣ?}&cg+=E> Co|Ѓ#l/%!6B̎|Z5ge:+ր8Ŵ W6)@IV<(r ZGym9RU͟˭WY']OSxV:x=JR2 V9w%f"x KWRHw#<.0uv)|STԎʆ!hfx}D`\"ތ]!-M+\}=rv,z'[~zVX BO ֣']R7LX b\[hIGr~"WH{ QjUJЍ[^BY@@}˥ejXQ yXG#%% qJz> ݔ~UcUe̦džUeŽV!jh%T:?oW  ΂ Q* wDH(˹d4%@ .[;3O(Lx"R/'@)iQj=+VB\C|+eT5]NZ(W^sa+UBZm.WHiI'W-+I5##B`a+=uz6]t`J1½rjo pDZ^H)^m+͘`#B`ɽ+DZ͛.W@)䪅red'g4 "W@K:xr3Nugq7 Ə]%FES-Oi=kN=vUh"C=J0gא+E/4̺/JףUꂛh2j 's]n<jJ/7nqIP:.23mdjRq cLeg䳽ӑ_f3ڇP%a⁆ +Ow&wŠl2sӞ=L1|FWSQph"cɉBQ%fDTg;D10DH)YKM9:Ә'}PUE/<6 ͥp#3WBV6h*ی+~_o67n,=3員%WC03EA~d~Nq+/xIz~ZlEWt/_vr}{7+߲*N< /u. [/@|X<ŏ&;Xȏũ} k= U|s^myv>=|Dex OO׳`>{޲9 _oJy'q'|A=?R}^|J-pYjCXM4t]H4MCޞVI*bRAb$i!3f`.2傩 _H42<2ba|YDBJGQ<ײE^ r<6-=)j.^"KՁME N,\Aa"ނ|Hnuui{ɽjF"3Зh2C6z=yIӳ$d lj7PZr;k QQ L)2b&ʈ[ŚArM"QWϫskPW5Բ>UOS|Uq=rO,v>[.R@uG3hXx8|nPD!j< ^//!K()|RƧ,1Bf2'?X-8:E~~o}80MĦF4q%NKD嬌v)bo4NGLJEKey횡螡R_-Ѽ^($hpa>,xSnow}n~>}̎fm4Tas쬣YG7Jwo7+#ixi~G0b2% .N( OoBL][s+S~ɥ<2wʕxdl7ĥUYdxW?EDL΀@Cwԣ8|Rh,9GL|*ZT*!0gEu TC$Z!dFud[`'& )fWWqSCǮvf6WBM];^m5ﱵu$ƺioRF3pA Ee;$DɅ (сT $ݩ{"d@cFƭFxH:E *xx, Uv,]Y_8[{uxMmbð} h󳯷)*r]q,Ʋ"jJC Ac2ѿM%Bz!yI`%J˔(щē@%'mPrql%cPEKe*FAmǏlJ+FFkM,Tvڨ~y/w")6F)D; &80uh"!6,%cu3rv4(7B馍`C,v|t=UFMr!EVjԏ͓h)2JOΑ|z:$'zPAb9BE$F2 MR]#cglGtΰ3 9twCaႢlY֌_'Yovr=ͮɯ~&/ $KaC- gA+H:f4dcJaiMN *mG-(7|L+1bwFvvAX11;cAm8I@fV"'$2&CCMfQoAY@#N+)Sχv?_F|qqR6uLθd_\"; 7qi1,bc "H 5"jttl9pcθc_<!ai\OZ#z6s\!DяK;~~+Ek%,;6dZ&J/#/9 4IR&j8Z:SH.N?:\~rߏgY!YAA`H4*5Ih'7UcF* 5X-Jֹgt\>=b7 &ȏ/*qZ\ks#i[[bs7Y)»e3|jyr0'"XFs4. dE $|TD$(q+&;9Ͼ3O/ zZ<_-)%J:ZsATGѦQZ-M4Wdz'3_5uꉑKREƑ@rsY$U@y:Q-gKB II;c$-zSh]. m#^JQז]m'K]x.F9_KW+ \+d!T;gGK3D T#`0OhZgFh&} %Xɴ.A(Ǖ?׫dqr=[ ةۄ쬿ly G~_eeyfPwձ `*LF ~qdqћ$TG,.C~佌?1$v2'yKCz󉯳I,z*4i2FxPxyiC魞rGX]ۂšf-aa^?B l˱,O*cdUjawIͣ q*.qsb:RyΗrpφ^ݭzbtg"?^ik0E9궗f4Dvy{ᠨSWu~6`)ףf^g&uSL}":jie1Wb1kƆ'^p:L@-2afKtx7-+5il'?#APoN pͦK7rO>̏ab&]\ @Qcv7O;6Lyt g&Ig#LPj;{oIXtmn׹I&1ft^ohn7V]ܛwu.'/;^f+''6/<" h+${$iz3Cmx֓s=[hy#†N^?ԗ_usǪ|-l<(]Q{*fjDzH{yֱKoݫhԁ?^"C RvKQŔs_~V|:( !)A%)IīRKK "cR ;]a58 ;5c*-:`B c\H)Ef*Xelu/xdoy|sz)ȓK? (isBR'e T| r$8V8l26ZHm]EZW6g]5Z _`U,`ofj{"W`Rm׹fiEײoo/>DWR;G=..w+˔1 8bt 7K˽ C_u0nސ {RX B+ILΆbE 7Cj~\Ib)#"Irdna2>uLyw].;);95yfM2[dMr?@vN(`م 'ߟ$^d xulגDˑ7w S S= arrlT"O7o7Dvp9w??݇sߞx :F Y +p_~z;CwZC3wӧn&|q [ƽ+>ŸA󋵵 >~fM|鼞T?yijUJWA('M:7)|EE_nSũ\RB!uO0Zbk~( %ЯPP G>9ν^` %kP&MiPbg͵VDL!}8y0ǼߎvI*";W Z&n-uZwin6.DJ{fc稒;=Awz!;G*5z״F:0=M Fww&PU$^Nǣ~ %;W&i Q^kjE4Sκ}=eYwN9[C'4Ӆ^Y˴s1sQq3&H e0¡buڵe"Yu*"2.K,TuuQ|0 6Ίq+^^}n>e0]4TuBB2{k86q`mةfxLzMپ|/{}mO0OREdS9[>m<1V TO[6?;-~dZ{?gd}G@KUS4P_~Wk_LP]:(g߅RjWKEJgѭM6zTj.Sn>\!h/UȝlUY)xFj2LSfT)YXzvQW|h[ oʶ1TgZp >y|EF<.q2*mr)1 oF#OCi螗f_*[并u۳܎._-Μ{9cwMvzcۗ^0j[L $PJxolon88-Րw^XQ ZJ0&O"!SH 4?B2ZrΑ3ڹ+9cٚdER7Vq͚FUj/ 8?/5},c:-n7|i鉦GҞ ɝem.Vk뵍%Ȝb͕SYM"zj͒mUR; dЈƜH"ZfWPc`@FeGևMd"^n9=i7v߫ӝVsWᆛwo >/.hyϦ a[>;X3n^w;ڢXHt o/Oi򷫻շy㦛kpJ?򆸲Õ]˝//u0?;~ooWvY]bn6[KlV?zǍ32npԷXmBs]Z gL[8wcz 9K;Rm͟(7³:'v] :frZz/Q75RNJK2JR^u3[G)}vgW5:.ʴӪ{0(&)*Iɔq-IӮpȶM9eR\J[m!3߄9 L8}ϓqz^9Ӷ&W*^rHA hŕdkdhQES';yzS)YUAH[Dwpt^6WdarE2<,zݘaS85UDo}G",PY҅fDy~'W>V"$R9 -3Ċ^kj a\6I==c:aO>%EOdu}x^<4M>tvZN/]~kj :E^7Hg[ l'y^-jn]>TT1>gy60:ξ?^IJ0/1Z⯋va(%Ew[ڑmz8ow9[j&~b ,nA^%7⻛zhɼkW?VkC|LMOi~~@?*'c)-WicGm!׮-5'EEYj m~F,d?y16v.鏓oSbDQ^oeO+#6es<$tfYuGY.ʢ\td4q>-D JSe`VsM*KOH!SBIir ]a} m&΋Mn<2dXyZː֗HW$1FbjR{`c{hp9w v>{x1Ή+f\}0ttPZ1 +N]5 ]? NW %^ ]Yr+pe>Zj骡44 +gb{@tglkF]5vUCDW/ю{s8uP誡彟jJTK+6!J`guP誡ߕVNF6Z=9]}`~juEp|djz2 &oK%+Vpָ}rmDW/\3+4H_gWd-f},v$vUEUuf{/l~v46O8ϘNNj~5Vit}8>YT,}9},vNxLl]i~Y^,a3;;-;拳Zԭb&DzX#l-ki]==''D~X.*oh;q9|[w~yv͟)VL-M]T;Jh"wUX+o-Zƺewx9dcQb~~R["DPm32*#-rPWVWYk`;+&kp;Lh5rxzC+F^y0tzy(t$~$(iMV:5GDyKr -:'%rX˻ zUNXɪR(5{DG`(IAfւ*;w:Q|X l K 2M6)>a PJI!(fLclx#X'iu\L:%IdLCR [ ]$zn.Hؐ HT5+ C(H%GY1n3T"8YKVzC L9&XrcFkVPQ\d8gж mguV(bݨaD)P,@!V6yrCPP\(AXލ [M&k U!R%B1% #̆ iBi A{-i`8FR{HUpjf@(3alTzCK!q2F0vPckF;jC5@\D\J8dR 3CuYkD[`xbd\~N%+%C9)4P@1 EVPp4j͒p oFbx@_1|!o:2CsRQr JH/K695byȢ=թB_ӲA<A; 0je1!= Ê/a Y+c/yk#& :w Tnj1".Ef*1' 2JʼnԤH &}ofElpp~sp&+WmsBGwF״̻y mS=ГQE1^,*4Q,]koG+vi v0&2I0WGiR%;A۔(eS#vxcDy֩stJ)h+>$ B+H_tLE;1ok=/SH39՜ $rA1!}ڄLkfƣv4zM՗Ud 2inu[ q{ 7lz,:S% 9ՠ1+ѐZDYьmC6YK1Z1ة`U0ޛo!_=N$ 1P _Cޅ\c=$pm SA>ᡔ(a3m>%dB;뒄 X "+P(v75$B;Ռ6awe{}696 HDhR:f4ǮfmBg$1S(ͮɀ!JP CFdU9\J> ;a!dً$>flq$cC:QbզX5Zl?XI3;YTg#MGhdf%Ej*5^ZzVE,~#XHh04mU($`F}WMCjihoK%7h'yh4vM ƤiB_nn Ar2#I* wCDyECrh[S9]nDVCqnD{NQfuHvP)HH Ш3Fi SRxEַQaKl+TOˉX1( "Nd On ACN'y/fȅIP(EiQAR܌EEH,nCɗ Ug&dd, 6#TK.d?uN^nY%A*DESj-*;k*C-2mP jx3XAN;X©tltEjaӠ0^ 2)z.$č0Y2'ʝ +ROQ!,iJ U\ђFn5"q3XoCn1jEro ":b9\0:fAj\!#6SR,AWKBo$_FY~B:COW4"革y0B.8! R G^Y?]˵ZS%v1g[Af%|nCrwU$0|_]: 7hX2ذ6>;o_tͽ_[*"Ќ  9W vvH?v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';N d9DPfWDZ8s"ѱ@*N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';4-Cpc(;R)v;v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; dG2Z='z5'~n@$ tN gb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vz?ޫ'/~RS*/^Pw]ZI_.`VzL%qR37.el\:Mo >a+ܨXZtE(`:@r^+cGDW ]\mBW}+BX]"]yP#+lhZ="nABX]"]7#+Zu{ JbƏi O2HpǣO虮4te?秫[埙ky};殶C앆v LWmzob\;<( u?OnP~=|dXc,zꔙ{bГo_~1_.)0D54lُ/w?}uy6 *p,ҬoC;Ӣ:y5G墛' v2З/n]"ongeʞuo/-:wJ-o\w\\rnVWpūszMP]~s~櫗/|С٤:rwFХotO ĈtBh5@pMn n 6̺ *JE 0"~4t=9jP-{`:"GӬ GFCWw~fwɍkMOۍu[Jl6/)kbg|qyVn-崕WZZg^y7]^O-rͪf'ccfØ]L6ǮWzloIѫn/jTI!^!ݞiVm3Z.zkJEY}nuUt|EΝIu1վS"ע"%nDk+G~vh}DPω&rZ)#DWٗxn;Z#iC+c&?b >f;;]}[3tЕU@htO2Hhޫ+K_tLWHWNKkÈ T+ڍF5S@)yAҕl ]\9%V}+Bi UUA+Xڽ_I(]`:@FZmGDW8:)BWV=]J\fmjzeBnVw>?;n F/&&^nzwg][VxӎpB}Pz9W@{tQ5qA?WҿV୿] zW<(" ס}?"˛λX p4=g.i< }~J|o¾C5l]M_zVw쉎,⪭m]zٟ} kx[ՕY7?9>zi{bH>NǾ8GRjo}1Wg wBI kp^wdR oZ7r޽1GC҆^&e ͒)m*:VEomFz;clQ,c*gw^^f eY?Z,{W^4%zbwW=&hmSf3 'u+{=^ӆo:~9/-#ֹ̛tq NjV [iڸ{aư(./{NٛrW,0|:FFu$]8[-Zι)n[pw^?~m;'Qb }DsY.d^PL%II)%?o[>_M8BB_;ZQA׾U5tKLEB THU~z-C`mLMUlR+TR,??8{.Τ;qHo͉ɣ MYU,,:aR.Mb.Y*-CRz+jp%1L"X\7JBVTg8B)c38׋ezНϧgnqmދ &i6(ʛ]=xXSn .OVBAyK9C޴ cH10xEtV\OEa;+m2&UV3:XLN!t镟(E5;+;ApOٟ"dQ(ѦhnK }q^d7a |@lk nR8"]/n"6XJ }*RUFфcȭSj5Jr@j]J(%&;$r$#+Ub1 vI7Isۣ>vMW}ЍS~M>0Ӂ8B[9m(Oy ?gsSb*h^eAg0]{iub>o|cji '˻h2Ck:U&ܼQZLEMFͷIf eofF ogкtNc@=m%!-FJZ=L?< p˶aj6;[o5FGc>vaRѕ ,&ual_b͑[K?=.N}z|+V&ߖAn3.nlYi=SZJJiN[."]lϺxفʲ[9kgjM]Х׋p}u{97FVc/sK%l=zs/myXRYE}M>RFe]3:T7޵6+翢\wXd,,d1v#H)$[feO/[MuWwW}X )u )Fzٓ@I("\㵩fߛlw3Roَ3O=r1f9tZ%r  PZG BҬF(R6=AD|sccԲm31d;q΃ =e'BJ`:٨Ju&F+q[F?udrSYvwpX|%fhNk5iAhuW>\B)6Q֪C yeux@(OC(Q7D}nʣdiO(ڜd |RȿJ&Eъ="{eQ|FI9R!8@ H$}EY,c4fH(wBK>?.ߎoy5HC5$x%9CRRAdUrتvEgʱm'N g.w YJ6+VD&j3qjuOyօﳹύm]o۴"z`]UlZU(ni_:ݲ65}NAIGo;+:yWom=؞3dR'pYEC"jUZb&ϥ 0'f xkr&&RMѐ-> j?CГ#E6<,XNT bfa}aj9`oS3_6ӌ#-.߀;Vfpy\l~|߼s֠Rmc#0(i'{i`Q% #TYE2 !fHPɞOR6]`%[ F .3%"KZd(湠v38Ԇvƃ@Rް& 3 (TgmI"X>O!lm@2{̢ʖ&UsY]Q$"l}5fÞY+0 "6ӏcQ6FD9  ;܄W1,=&M%JBeuBhᴖ6z/$IބXm@vHg \1|dQ3!HM6`UIW u fGį_u`\wK:Xg3-91.\ܥNe@>8ֺ38-r ʱS.I'#c#Gј9sfq,x 쮌sͿo~ |o9 OXd=O_-QjU"_5 i+52:'tWCw_ ͓'_O)?ff G;E6)ECs\gt(Ra-ιKXi˯@s[*0I,efoLs,QPyfJP9!Oɹ) d}\_-9f5nvoWigWqW%zq[ {y-{tredb,{d!gR1^k[t<< iR`AieY(6x2 D5eZ5>PP7}XJhq21=uEԲ̢\bYfTh/sJc>h 9!wѺIeΞ|oדzg뱃G>Mny75~_UOq€: ˓Jl?"R'$xQ KEKk?ZqCx&4@Vq!p*2la30h#rI1ǨAzW.zYÐ΍ HSu_<*;+#Ǜzk!]IFTH36:]ԐN0ňbofgٻl7և=h8TghPOrZ3_@Xl/jAߟtI-߼xUa\BLZ1T4**E $jԑ|1gjؾ3_WHYsQ"-&~~>K cR.zq>z#g=n׾SMedç4S/~xD&/t;y^{N ?VMs6q;8/QιmlKS5'oxPj۽lx _뤔hy"+Yͣ}Ye@!Ɋ5i6L%R( 2yIF,VدӒ[]'3g",iz:sm|99ݝ`g*rEZP]m򀖣Zw'10]sbKtێ h);mC*k "1Di,jd>)RS4ѡdrJְk&LRg{A=ԟT&h>? -5X Z (4.Pe )ڹ@"(B5D\t}\ D4Ӎ^_xX9<"J񤡈+q +vk?ΝElcc45)mX-+CmGbA\Nx;WdcQOD;a)fJA]&Yjmha$mbN k ٬HOGxi[NVozO2N n1dQ620I-1X '-jck\A`J&( U뺾^WKpoj?te Gݼeய]35ݬ&eїt~oT-)ʌ]mձgGϹt%L`MYԳ #$$UUgGU3 Zm6#& }r?kvݔ*YfqWyp$vXMYcBTmufKYHLNRq2H"V<'*\Zf$y|tz2uQRu( Z50r{bMih:#yQ.d]0t h60ʮ^Lг֥}gǔu4 j6Gɢq dJ/ߠAĈlzIrjFeKE"g9It{ ]7vka=!Ay ɓycOk䈲Տ/ZaQ܀/4zǣWZB#>q!8}/_ƿ'_[2웶?ҳ_ez8nEN/j/6>pvNKJ.2B7U板g^e'љN]Av` E^8ʴG>*Ah\b(RDe,1ieCg4ރa"ßfJpq5H<$֡[WL¸#K"C UZg@6g1 2[oi/w4} %R`i$8>E7ү ݛW\5+:wYJ)ЯW Z15:n:DԮHM=tt+RQi o#Un-v1ym!aJU0eR@cQ `&YUfD|ƽOW1yXIvn',vR 7qoC"CKr楇™]wbUkM *m~=4e[v.,~$K<P5w/lDO;~z(F5&iM _Ca{gXSxN;Rv(U("Sr }Szۓ=Jz`|]/{"o>O)E24E3H%eѡI`љ!f2z%4U>=V&EVѲW#`.]!l-"{6l`E< jBUygr)^XQ:-k@1WH7dEJBFmϿ05tPsb'E;vp>&(BSEU H'Y;2gښq_Tq^TC6y n7+.+yH@VT c-~8rY:ˣٜQ٦{5KNK%F?*&WI.WKZa9ZL4ؠ?VW`h}}>kIJ?/߭יxR-8сJ1W+r]SS (Ί:Bt l9|\Szh^u}Iub$u>u_<,Vث]sV~>ocjޞjqGidlHjHR=aaa7{bĘƓVF U,'d<\.jpmrGlInu\Y:+9i#e`E/i(]H~AxG0,nznɱҩ[:մ'l8^< m??{˯|ow?ǏoSOJ&~#{FUkho9@Ug-Z^sƽ>ոA; e !> ~?fBĎk/]թK̍ .Bd.BEK_nR%*V S6ߕKo¿6V{#4kpCYCT Jc%D7 )A2rL`=K״}lRkooK(,Mi-t&U_):;N[mO.:-ܳ\|Vw1jƤKzU4[h-J>G UL<'_Vka|Z(G"sL^:ݍ 3 ҥE`@SiAQ(4}ѻ.j6] B}矿^HE߱!1 iOkp {5tvܧO/[wxx2M].՞Wmzt 9y:m ;jm%;fwtݝw=9}DK=ۛȭJQk5ghۻ.~AB.FiԺKfz{َNWmvl; lݺۖVWwweZa:lot;jyRR_կ$ڬxW[s9e}=;Č6y?sWu:aw[niEJ]Q^is1WS&-?oildrd  zރ=-8lΜ"7HVL$9s)m L1`IsDQȪ)1JșSkE4exP2 xv\]69SaI6ؠR/2)=5ŸUVy<U ޚvB/E ИU+A(,f.ѷT=VaN{{!lZC~SU{ oȾ 0=W>T@ZT62{XnS9\_ZJ7w/E jdHPV( SĄkGĺlD#(-BV,^ZCW,@J)HXE]!cJLA)G&TV[[MFpy?xh#r ГZ,R!Ee^ɤ3L.R«G +s[9倒:Z{Pn@c#4\b1ɒ]͔T? f*CbN0P4<ŮL1"C&;-Oe8oUǤ|u-~tK e5@6d")+JRۂuL$K 3҉Ǧiz."C3 sJԻ(`k8鯮cN;N )}5<“h3kdm5d|ܹ/ eFZ}CUs[KI2SA>UF0dH`bΣmv/tj<Mlvz5'ҴS<#(n"~6P\@qġCqB.Tae3j c._+X2 F(j<$`I{h>OS?<,4ـaYt6-Je >j%I0e8hCz].T+ ºT*{+o=~ZJ8 "b{Cц\;9HB7$!EBo5}]{O|4KW`ו"~"Xz4 }[oӆFyq`JR:Fʢ}&tײ+OF>is-1 @Rx .4dMI\A*1"֍H|0 BvUh e,ZD`{(RX= 13ڏgegm>o;>?oJ1Y=b5D&s  c1zϤPX&Rz0n O g GF`cAю=(ji%")FQU0YKJf`/nF+bNr΃dA$Pg5r(M?$ $;JQ^H'Yu49.[+;n o6[&50EGp-k%ESbr!f0+Bt.*#Zg2^l=yqXoj0 H(,GBʖxl6MxŌ$T\Ǖ}=dvu,ĔP._MFpJ(5[]MGjцɸoczqJ2x0\$(x? N}Nlx%fvt>q%h`8oCf=wrQmR׬*1r]{oɑ*#!ld3ֻ!Sb,2IWݯz|"")+[fwuOW뱆XثUs 5|&>C,J?uJ_)Y+}~=m/6_kHrgE_2kmsyΧzʚL_NV=id48{i2خK;oo8g%`Y:]9]bs8ᇷin4U 9>nx4~xPu5M4?&6|\{6ޠx3`߬)mZ.jj.[:uG+cm3M}WG { TTyS !1XX&4-hkIzLUJ7Hh;JA&:uYOkU=w&CUU޻kwMwV+hز>ZYKS  0GM,GX5p{Ψjgoc/$|m_ZE+&^h;fG* ghr?Oy^IspOo !W(^{Q}T6yy>ymFn",E>z |xn}*}|3 +#7?lD-!hPq3*tMA(()̧^J|V>sLt.i Ϗ^T>( +|ξ䠅$BxC󸽹my`󿕋^ PNGz;xS{b:ĹR氃m$92lo٦57_#@龻E9A!v=M[XVEr-ҟ #JvN3⸖yH .pRtWXhoH&M4u+¥& xS&aJ;f2Rʃ$^t/|%Z#gESd%(v$Jz.V~+PAC[W)J^\I+-sJ:夡yθ\ZyEfZ9a:$%|F}Hu`r~G쓪a HitL*"Q* JJb,#Fc`^~(~f"h̺{ @H=.@2d.أ9yCkmNﺣ鱎}q)i&6㈸ 2U 쎀Q zfPCH2Z!Sp rM #QY3Y,6ZF(@&H@CM-u>džc !ugwKS L0_ [Ua@d>I3 :M6$&Ia4;%y E"1L&s(1V $[zkt|!wf4 h Bqy[   /Kx*,&3&[X;/EV8Y&3Y?|-.N>Z%XٗoU$:==>M СNnV0dqNnarg%!a8n9L~TV~-u?O^^݌/fbul~U\[k;c({3.. .MW!4o#}n4#,BJ2jPfK'u[/h:fsolrQ Z>j=ɦQ dQ 59l.ur b.N1wr,u*N_zqzחO>,A8FP[ ߠ򣼌EǸ/'q10 f1grH8+0'5n5,:C01 ZGK!zƌg>OLj9+h4Ĵq=cSs3r&>{87sU-۬kz;zs[zKmQm9w&矆mOYi@n!%OpXt`VD ]E@QDIUF{jXiii$FĕMe j`:=Vʁ,pN;[B7zf!% 3Bs4lk2r $O03;.u30?pB G˪>zȋ Mo# Y6`J0; b:#JFt;ɺF4=h)meR.fΕ"Yn^.m!J>֤Y1Ƥ7D pc`2g$@ ,Hq.<8vQ H.'Q GLEx bňHP<`e \o2pK2VE{(R{Iy@` qht^[nn[󿕋E4o_9{ 㲩ռS}H!z5pLS Auqbf]oGWKpݮ[~H z{,^(!)N(RCqZytuUMWw_4ý3q?/WfPMx3g0@f8̧pj<{}v:ökS7AØ(@+9) ^ 6q=QxUg[o4)y /}k #7uגfi,0<'VKd0'e~ZiPG޲l+J\#vy![lU+=\JJhs&YXwpϘzw2brS ƣ.ƓC6ݓCM[4zrqC$FCWYEWVɦ4+0iTfϐX]\uh5k:]!ҺٖN1\Ɉ .ױU+DihKW'HWR#+l⡫"_(ih #QŮ0Lm4tpj<]aNՌBVECWS ]!ZxB^]يgfV:v};vB+2hC)FWv -]*z ieDtg t(mJq#0djk6p0j` Lȇ)NX@)@NqFи-67vq&Eg}xyq܆\nt,sR$y(*DRn[o*U`Jz-rU.b=O;t~f]tlyАywݹ2|M߫IIn9 R1Q1Zj*(Т*V<;߫!,<{_p|L=w`չ?􍶌򍟳_; ҡjYrF%E;{?[Z]6״Gmg |J~ڐ3ѼA!""iq-m,tXDtT2W.< -LX3!\bS3!JZ}&N14B&D5%5-] ] JG?e;BWV4lJRN00#Ɱ8R4ՍwerUZzO]`!㱮fy=֎ۡS+ icrs.EW׈X ЊǮ%kCHW1o8h5%2B ,-]]Y)vM<$P;e+DUKW/(&zJ΀$=zʖ:i/[qCٴ Tٳ-]$zUҮUqpT6K &43uL$gCRS,٤Wtmo<ϮM/&*ze+eR0*RxZ KKm\gRo;[hqhUc^a$|uF6R20&|pVJ7Z~z34XQOI"Uz|,RaLy|[a쩿L}5Sqqi3S.|@^N7vy)(|Vdn-#G'`k7:rHR.x% !'%RڔsJVCLnn'UL&'jz}0qZ?Y*AaP /gFPv[y{O}۲zl"g2l6Hݙ7IO@(@1? JԻV>a}y+騕՚Zz$ue 3k/1DPHm rk]C7~I+V蘂2MprKp-QJ'O08I5ץyN!Hܨ,>J&? B̃?rmսu=iVQpcdˮ)Ɇ:ˮ|iI`U*Q`H nKiey>楗]7Iβ{`Y -Y>55nk (ж TB /Ѽh SIOST ٦=q1ZpMl'ţ f];_-7-a>\@yU~Kɬ8ig'gC8uyM"@g07XA 庮?KW{t]jWj^REQpI`kUPE͡ThpUٽK׫=gaU*yAz?5ۤ}*aKu /:TXm6ܦݖrlTy0uMcJa!a%XR?*t]tOn\F+>ČVF[w!J>BVr-Ba:YJҷ'}%Vh##o%ZČKHs~bk+,i4tp8OWMKWCWBЃgJWx!gt/B6~ZW-] ]I-ɺBb+ϒaDkuJ"NP x_AJBWH^mҶS+-153pE4E!Z)NWRNd/"h P-ǮmJ薀,Bt(Ue(zzF5lٱCGCj;a[-芶tL g j ]ZAD Qr)ҕI 24CUq?3\0lLۢaI%_&\cNn6I$cy \i%6d*ZhM%y.\"XjB!OZ؃--xd$p= R X/6^y)9ϳTr)8s+*RkHpQ5B qOwG͔Ed`x46kUA Z;\Kbr09CBBWV6>(UK:E2c#+̟&hk:]!RaKWCWRsEcRr\ ]!Zx Q磡/5$wm,NᒌF{.LEbsc3+&7|lxe}8-;aƼźVyޗCY/x.LUa.eA߿qSɯ_y/ͣ.%% l+{wvuԗPG]WaϏJ;qzXCn}M:nrW}) b-{ﻇ곆7)7U]rJ38hN1 eGoh;oh;a9[l 㟵DτUN2scVfRd jQ>|M䊿_WnzTɢ羾ŸJ7p |9g\s?痢k:n$d4WUiv,X_k@Z[mlҥ/ eE]w:YMB0%$4x&WIN,KWJp^2!r93Ox..Syf gygHQ垞E1ίMfzoq,''wOn݅M| 7O~[+[5^xWBx>-@ɽ~?;~_nƓmW6cإ{, P9 .ݹA8#lY=c{ 6Q2-K(#nK5͜uLpnp2}n `*sF:tY{ :zyم,*?_#1 i Z(vgFgA?XJ\ާ^ A5P٢"Z=x&/gK ),RRE\tA@AWhrq~N>#^X-8ƪ;J&Xд?)vƽKygIER b+x Z6O[EKm-u>rن $L@J3paB>^HIEsmDʏӼk'yb1D&i0.IȠ9eGƣjkr~.Uv?{l73ǽd9f̦DP!$ R0a2g<Z 8qfy*CsJ3"Ro,PGs-zE>Yn2Rn4 5JiF~RY>b0g_h~3Q$h$`f|CQa?$ܶk6C]089ё#kQ fdξ3Vt(>WZD1\bx3tn<:Lw')ALcar J+6i[Ɂ9J0}Ne3OZg`292+6&^`#é#d{4n8}>n 6hӖ]Cm=zj|{n<7[憦&sJk\z /,-2 A}CJ2rEU3KeN*XJ3gC0ΘɃV)@򌸚568OkZ;8popE_w)Yw+.:@}9&6_zG,jCI6%U$m %qReRHaR8eYsK#ȡMs(jCJ/tE~e[I0= cE݂50.Ij93Ĥ|*(O}R!e9SR:efs68G)A\bqmwKEU03K8"\ /LŧC4ϵ"2R\ -#Vk ,p]R2JP͸RSfX5|, $ }c@9A5P'EVcף~աZhD%H+Øy"( dbh/lhHU"jRnyNdBsֆyB" o!ĬN+*ʌ̧4%F:9e@=!3 R)\$U73ifV k;rrkP\Dmޠe|fNK|ox/ld}H\?M/%^*-.XWҲ4F; YKsR3H]oWH;"{ v&o0Iv>d Hbw밭efE*X%S#uV* C2'Zh$ؤ` fh/#vЂ0хT-#vk܍vP%PP58G>j8#HIAʂ3-Mf@oY#NI  DY?#ȁQiL5rҨ|QχǶHZFDrD#"sC:1sQJn BiΉB)8<ģ(1HQFXlBqb%#1)DXҠK:Y`|2Gc-#bk܍_.L\p^gk\-.qq8, c5psVgW!zڧ:ae5. !ovxZ]j< yn8np8QX/&.I8ˮ H0vӥ}BUPa*:f_l4TE=LpXᵌʀS)尼;޺&quwu$/!*H22ܣ[. H<cP%b ŀA' .:0uڀI@riM8ے-Y.JκWL>=۰Q?'rM?{QkE wQ7$з;}aUo~J97)]ƓpM\ ^ud~oV+{7]puT5y3 |%~/_R`3eqţ 0@%2Y}^t6m.zv9.;2c- kzv;kIEpr0ɿzMY%ʩ?$Rg`׵u;zRXvW.oVUM߃$;|'7WSCM4Ӕ9(}-HJ{&#tF֬20}Pw)u9se*̊3fH'?5Λy}K ?v{ugO \CCZfXVR([:cҸ,;M&\Q~y:|ߧQW!-p|,/:zȺigϟgzr;q͵t*T-/t6|!^NGeWt:L> Oeчo^vʿo;n"n-AmO7}SbNi' gzKl 4ٹ @`b<[H3^Dan_S/0DqZVp]:8XE{a&+VY,z;ZIF|"q ްd`2K- ljcwH^W\>g{kÀ"avξY,R2A,XTM1Ի A\,&ɟԸR$IS%YO&-@v{vv9bM⯛%v;Q{S>dkf ŞIw ojޑ̯'t[_;+wyz:c9EAu=T߂]k}ǍyePo&uWlkio44ѭ,7ﴻ۾ߛhc~=r3KgMr(eiƥLPQag"†D$ [-)tQK,B*ы9q%<sQ*0 Ibgbs #ED)sĖ#KGov@/qy}(AHZ]G a(!(`/yD̓۬ݺg'pcv5@gQVSNz갿ll o_ːflbiO`}{ %V9c<٩3>xg0Y. &cѰwG; t<r)3tny$<%V,Kɏ_mZO擬Y H(򹖜N)bxЄꣅ{Mrɠoz/ } arr2WB)D0!a |0n5b{ Ϸ-ɣ\ 50)B`vHbA¥+#Rb#(frV]_uGFc; Bx TFH m8/wҺ*~a *??7>y[oO4`J mvI`Vͫ]T(IsE6WEFƣ}ZqX}{!M*B ZGюGF>հSyaxXi}iF3oSNC @0($P1,~Pz%6!`BL1jveI4aMVmgj 9Jwp1[M\K+|phdUM$})}HsI 0-j4SX|Un))锓JC d93"sB+(L+' sYߨ!s@K?̍:8OTƨwD@4:&(fESs%1  AyAw_y, { ~zha݄;#5U7R .P {1GK/8#?Cp)()\8#*HTY'&#@̠  +H[]I[5]pG=͚ %b TT,!W_JX+?-hT>xgvjHr<(W/]&:Wl:jS_AIetf2>Jp# lP±Tk*zE8sW޼}]=i F`, 'M$,$yM+Vެi&Mz7hW6mH- ڞQN}}QK'WۉO:tZ`ҕ𙟽R'LY>L;;ٷi K;8+6u#YܻeQN 9w$m3,ёiC/ C!fJ68R$#Y@SJx-(l^v#a\R}ntt{V*R/&P.)%qe߆A.q޽7M}57OKѠ$Wm"k@7䚥ԙ:7zU jjI&AuZFKQz j)Typm> Ѭ_(П sZD)3ln%9 ,Gj-ʏ}vsڿ168JL9*"s*fZp,$h 'O vlڅ~J=ڄ­RȰ,.saREb%ȹ{& iuϻо4rDۚt24[KMh"xT}HJ7gXE;kC9:jC?vR#c=ލVLLryå O=cFcPg>LLj9+kn#WExl4\uRهTOY$ʔT{$EQ(QH֒8}Ѝit+{gM3Jlڞe [ZT˹<[-0lrmsk~2dYEz7-c/y֖jju(y 5OܡS~.e+XтJMַT\h'\2hSwV5h$P O Qnj)Ǥ IԞwzߊ@*AT6CHTS.:bFyl)@,RV0  Ĵ߄w5o IO3v,mn]--_^}E6\-z ~qC3{]s0"4C[~vjgY֝^� 7]yI47VҺmޟ|ᛯ^p獖|6~W@.2/n-7o_}9eM{o`h^wA=[|nn_L=PW-M|s-0nu[>,58"kk_:u:0e\zn@2lh҉!#Q 9PIQ hPCJTv"z+KtZ>5!s̳^)A-g+OJd =;۽yپt1;=I :w'5N[jYSge˫j5zJHӹlL25uU48k)7*:&nɘUEH_ ǺrYR+ u o]T,Y9l z]r詇wz5sc%zQR #kPbgcZEL á`1Ic>nvN'% hRjDҎU}}(63] 5'h(7a 54=k*7Z58}9=b;YZ{@V>\f8o>=TK2C*.B1]N6Z.&<9FWbT*P]51V@NPKSu~ CwzHةk[[ ^v-kouWL~L=QAjGz,iZg*sϪT|vle4OerVߡAZ}T=gZ**zurZe8q:T2cbSů2p1GA&G-uYubo8-Vy)+@"\h"m]~>VgJjR7'O+tW+1(B)"-ׄO` LvCz4vChǡWPnBNv #+L#rs. ]5J] ]ֈ<"j ]5~4;悖୫RO;HWVѕ&]5z4tКo"789HWrfDt%5lpɏZNWLtuteXc+lxO_I7dff]d ŒbS E褈q KAL2Ukf1+%26&&]kb!`¤5l} H40GQ_ +yK![hV b+}g  j#ҤHACUJF Xv."-ZE&ZWв))~\M 4r˭^HiM VG(,c v1O\N XІwͭiUJ:JD <,%)@Vx*\Vg@xh[W ɇUV`pS .V(¯e2Z R zq-I[O}U,)V@ Qk Z/Q$'Ecp.1[1W}뺘LJơ68H2UU\ ^PC56J47#m,R8`.2,(Z-i6Y]r*)Z+eb$7:B(L'XQ%"ɮYXtPĶ>ZNlU>b۱)-G2W1pbbX2'J99!鋑!RE1sݪxNqXQą+4KE _Y:[5xEx?{Ƒe ᯛ,2?,` ɗzFBhR+RN9(&^JE[l&cQdtս[] F@Bҽ`tܣƬC5,ӊPѪ J(Qkj5г|8+¤T) 1vx X-V-tkuTiXYi)@fTQ+h \xL +Q57!:Ci*dL'BXp6PL:@@}\E[PB]Q[܁pYV ++5 e6 a[om1!$( ""*f"HMwآDu>UFߚ K@ZAAuT&t/Q*ACmj(hS1 E@(QGRl mB+mWݕ(ދֈjQUDI)be+%r5 y}W@Bb|M6y-Qd+--V Per1! aUhޣƻ=sA yrM+tQ̈KQEn"棎1͋B$%!pBDE!v>v0P7e^=}HWtx nUA׮d=2[˪t$MLc[oT^ҏ[0V%SdIW=$ B+X(IT<9hQΈ ڃVsH&5T^Ueh6!+*]xqy{'*APH܎m }KEW珪d!T? y E*Ίf lZA_֊ANE"Ct_ϧSyP'K =J2XktdDP(c ѣP.O9 8j&! %:]%@_>1&#TϺێN)ڽ,C)y*a3m>%tB;뒄 X}t kR([|Pm!ڣjFP,f,Z{6=^ Т td!.h]`QmB$1SZ(ͮɀ!JP CGdU=\J Ga!dEh3A68FtflNX)j^ ?hV QGH$5$YI$ehm@VӥKoU4^"{BZFwvAVmy6*a* 0dkJ@~-nreq9³4ViYr$YK`0u1vtrlѓХE4IvPp$J(umT]kM!JKBs$JVO ]z31i|0a#[UfĞT2z7$%xKT]ж+9]nDdh8h>jRʕ,]T*2heFCj0FQ-HOOdPJ;HqXUaKl+TOk XQW6b]MA53`Q|AWLF/LtB)j zf,(jFb1;$_A,+~x$mDVc-S{NWjtҢ:jՃ*M3| R5fҼLFALЁha3BBvZuZw *UP]xZ*;k*އY[tڠ@~ǣ&X©vltEkaJ3^F\]črro.Q{OQ(t %׆!I@ b~=q?XoC1jEr7 " J.YxЬY I负bc \- H_F]~B:CWW4"흩y0B.8TQ~Ջ6B^y']\zNॵb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v='Ї[Ux!^c\-q)(ظ,KfDtE ]\oBW6S+B3]}5te?r홈V+tu\韖Ca(M8-Еeԩq{'wnSh~M(1{>nW_3DϽ0n{"(!fr=_Z1SL3v LX5>ς3`wO?(K̛tubE<}{WsZ8!ͬ!@ $Lo ~nvy ˲jeφI`IҷlWj!NŻY!^{,~OՆ۠ڬlR+͓O ̏ th6v7F096wHOҎF7\5@h>u@(-Jq4tZرU'S+#;]a4tZZtE(wV]Y#+6v<Ȁk ]]\BW6S+ 9ҕF1]'΍wc+ OWrj ujDt~D + AB೤c+j< EW֞FB,WCW#ޝ ONW]W=RahKCiiѕ;էNT1"`'h>0Z:]Jt,*hc89{ iυg^+ rqg籗00\/ Xbl~_䛉_ee6_,/˝G\@6X32# i Q{p2`l z5ɵu =*ahls&} <ˏ*}bZ膘jQlעY8?!I[YrF65-5C A%"=+Rnyl'A#*#ON~Jg{h7""Q 3"Pjt h#GNn6:]\BW@Njg[cFCW׍$/#]9Ӧ)Ƴiʭt ՘ R;Y"!:]!]]Zm"ʎ9+BsG CWXƴv 5z,tEh9uWCW#ޟ [Ktuܧ6V?a(O-LW:`JZ ]BW@;T9U0JS5aDsϭ;ؽf{H |l:9I}|@"~5k`_ۧo;y&P?8Cj2-i6|a]\(0~~z/z=*|e]@x-C" 8UlR,U*iП#i[L֯E*QX>Nȟj=mGElS$}w9|f8r^_G' Z ׉`τ 1}hf߶)חK\MYUlV",0)&1uKVJb,[Q+D_:W.߀ߛ\Ej{r^L H zᯋ;/ǚExUv%꒏G|,U^[ AXY0Y㬪V#}b84/ךfi'6f<\ͧnq}KŚVPjWDU]}7j p]'5͛ݻ^x˂37s=? zy}} =_mu!֠>ݴ cΡl WIo}*Gr6(뽄JǬC,&B6WKhQE5GΕy8WHΔѓQFՅQ]eSC4m7 ӝcM9/HE㛰kM|NB)TM6b4So\!M/ܪ*Rm>hLBy ZAαSʱ#ҥRR7)'#*ĊuU*!Bj(B6&a}.5ib&iαʱVo*Uں/^o#:.M䝔_k/WOwM;M~jn? }U˟^S1 nj,zn2/f锖kb>u:ɳE5J0\6[^.ߝm!wLtCO^c~pjo#I~O]M X%"5~xI$ہV )RD5-ʞ(N3]SUWѾтa^{Jfp* }\g{$1շz8?w@?yxi!&|>ǚ@k܁T QIO>$vOa|S4F}upf=н@f_gXvGuߢQf0]~FAZ~B86?_F{|:z9q0n $tKZoͼm[ښjm_l;LL 430Bf'$@̦O(`hHDVF^V{T>'}|q mBO۫M\!bZVCR$D2y̅\y[I0ec'_8scc IJXT%3⩣|bii OS!BWk}B:O]vmmq՗LW4 e7чb0䣣D1f+KQZ㟵s44Udd1ʫ(*LSČ ؎ll rd}XGsr\Ri*c2J1k4*keq IAL׭fF?왑k$#y=P5`8Cx)DM$i}1?sA'S-:PQ8d2oP"E-?@l#Y/m]j6+q_8#GF8￙ =CyyHis/dm o?Y-"w* HP9d'`:+OF-TSFVy'V ikSBr9-2qo?Smqx՚}DyߛԶyLw<ZGB\=cDGrޱh6ǻӅL(nv <$ F$Q[Pjx/Ӎ ۄ)oc6f"{X/Zxt/x{ J8;-,A\v/]TW̫[L6ּ;]S}?ӹ=Ʒyc NqP3'Pf6'c>.L(((x`:Zm,D%De8FƋrUJ~}ՋN1+"Z2TIuyEP.%qFT%dt:[C[ jQ^C,tYIv[:W3gBS,6v:\%(_$l*cRĕRz$K\*/Pj]zw?u4뺨9TƽrvWًbuM [%8p eNFD0L)aION-A Up[%x3pq>q :16j2VmLg2aҵrQIpDbJVF!Re,@%D.VN )@F),] ;2Ϟ|g7!hƌ# !Qm-9IhF T}<, ;L痶~&5a]Z0\d43[v)vhرgbGFXb'K)@ IiI*xpp=Q:A+Gվjߞjۘh!h$"r%tI!Q$Z!!e1#ܥ.4+Mzћ3@Syh/3[MOhcmIHg}}$c][N]()y.Q&FG^cj7T&4Q,RuE@HRF[T-ņ_7b+]Ͼ٦]I;Y,KN4i7:;1hNt UL~7 JbRH+VkYyB-[o-O͠t ʃRޠ!Q2dP.%C($" Z$ZXZ; Q^Ҳi)p(6F%18A :gDB0:Q "8,Sh<Р ٧GTb)fO"0jNΣpK$M_K_hY&*f\T9|=RٯbZo!-ӜNbTP\2ͧ'6(iVN<(L|rc{`=ynf rj#YQӔvHDdBsk`ژR&|…YbZqM(EF:!}r@H2ǃ q0[J A42~d,Uaa18 -Wo9\iۻ}vdoTv8 Wؒ $+iCFIHo M\HDr|(bsIcƞ p_2lr4tpJL m;h¨c"^Ca.6ƣbb jSAmڡv`w'U(H8T|qd*5u8ЎWsТxPw @ !hE{5bMr%#1 ߰.$G:j[zOpam/,d`<D,&""+Cwƹ}Y$$@p-%D#%N){DI&0dʖ\7@ ABqS~[Ђ3.8"GK:[h| Ϝp#b]:٢):Iɡ( "pŝ u%{"GOIwܠCLAŴA #bbŤPpJFRT<׽WZn٪x+puq9ݞ>Dž8 R%&m+o&}<v7*Jb:yuO`Bp:pR_"jn+k]}aEC`KtкK6T^1OEJ`0mzZfDHږ[Dc!X"`U2h.y/"U6:H̤^%xK\&.%\hh8Ax\&CLjAu[&gEG8oMm7#~(`iGa5O//b1FŢ@>^|xv[rtװβ\(LЎ gJ!mL0X5pj5 rZ33O9If[Vi-yBgxTfcёhJ5̈`"$%-nX( Uѷa5Yd]e+* Nȶ D#|'NRH2CS5@h tKs-ML$" H BZ"D3YbHg&@2uމ'Lz5E̔bdqk eAdxI"SnA1 RöLox<~&:lwD54Uh_c4)Y GKQiIz͓=0'La|B {V^r)2̿X{w!!x|Q&qk K{OQnNF ]384~8rq7G+ tnɻF,t_ $yǐ&= W &fp 4H?Nb/(J+O0k.sAܩ;nz$I+A˾4vw|0GaTP4&%voװV|Dǿ(U?4Iݫ6#XnTy[y0vp_|?X-m;QCmUi\< ^ B$$%],+-/F//f~`Xr "ʹǷe/WUХ2xJ.+uY]a =+u'i:G` Q_;Ց=L}$Sc*z ˏ0wO߼|?7O)3hzk>Oئ __h#*ʛ͍XZk=iʵ*/)D-~C/ q@ꛩHyNfu8 b/Na_2I%E%E.bI!deo4nk\Ց,!%uB؁+JF  ZGKsvdgߌEm,vN:E)H qx){l xT'^GN oTTR QY8Ǽ) /p&d _I麃zv) Mєo4xKSKw*KU ~l$̽2ƳΕ:$J$ymԒCXX-EaN&˄A]Y}N3|*nh<; "yRIծA#'jyх(Џ.H/J+{ Sh4 Wpg;ar0^^TVߘՋ[7~)t1lXks? az|b\vp-ݷ[5TTe%]()CCAK8bL|RxqR S/^NxP$ V܏B{hN. `VK Gȳfy9] G IQW׫2?0^X߹ @?U+^Aڻ;ӳz0{7@3>MOx]4zg,b} q΁Al-V^G|gU ^NBߛ|#M 9yel)uNPkʃό2!JVI4(; k mHq.%e#I6 F訧2^_&Db ñ I%:iM$kM]LPKzنfA5\BiQ R g($X@IND-gNkbɩNEC T܊޹y MzJlQouܒ{l{ۘƜFӝI.tOϣ?~5lϥ,O(SxaL;>$|:fNit9EZj  9s ~GGJEbUL1cb6 _LC\A$Ȉ0Km{@'9S0eT ԙXvPdmpiGfͦ&??,Ëb msҡ x֘'3>7dCRE9'2y65 !e)h]=ǚ^η| Bs)o?ҽ*c4No>qt颺 }^]ԹZ?rz=OzJD7 bT[;[ޫ3 Ds[I7WK"m,va\6-ۈQۈrs>mm+,4i ]!\#BWVtBC+tjgEte++H[ *tBvtutFd5tpۣ "ZaNWRvtutl+8o ]!\B5] ]YiMX•-thu+DiYGW_ ]-{[gZ.ztZ5Q6L4kЕjӮ0[DWD-M+Daҕj,hXl1Kh8$IYic6 \u+y" 'ϯ_Fë*i<ι-1zB;*/&oo^oKȜU(){o,SzHetW$|pjǶU"^04 飢^Yζ{Ag0Dys2lnXƄt)'jUFՁ&@%4/ĎsB__)rݷVYU'sIN7`Y/˜\@l᥷ lː{˻ ;Y6s *&\ 7_q!irhrwBi&(:Mp49n8eEtq--o>9+a:mKWEWRB6] ]Ic-+Ly{j2hk;pJA+"g HWv6)pJ[I i ]!ZxeQnXL sBJi2(]}5teWz{Dt[Vj_޳2ZP Ұ˪te;ڴ)Z5tp m ]!]U=]!JAҕw?5ɃFw8+\>?9}HP>۲Q {?I+xPܽw޶AM.VJv ;]ɲGY%rp];$gCg׏$joYx Uf8 Zs^7ƋwP }|@ki{ClQkU۳8R;O9ײ{ fI??G2rFڷZs?2M+)rEQZra] JbzPu[LY%laR$5'\0o\iBx)xIuѾ|ᚭ, ,/x:i (D*3G FF\h0id=~s?ey{+)LEN*DstQJPN'E-p Fk֟:DP D`,PGQ [1 VzCdupT}sTcE{{w.z$>ƴvz*umt>PUhx򎾐w$ ey2iY\dei*vӾ4a; ;ۏr0n.q[:7؜!)-I̮0@N&u'6wgܞj=#ڞꌱ~)Y'QiŴ4po qKd!Q Nk .$dxyfD%txݟ"8L׭6nn"}jSMv$en@m`|2pJkk8AEoPe 4%EԂRΪQp&Z 8VP``&EGuN;4ulۂY_zմ,gi:ݽ'k_ߞ \M ]s \:R'6m r^qZB aJ%S7*2"_BIT )q- &qK#=x B)qJ!P uy4aShESƣ C=GOE _s跠- 3S5Bu&=H*!gHhQ#քln,2Fb^p>$ u'ܙ8&uN!f`<DL?ED1""qkqB̉ Mf;B$R2BgZIyX ISfC8JM `)**y4Ȥ3D) 'w;g7" "G\ָi3-EǸOxmϑdE >e%EG716䗺8ϬIN [\<0٤#hFI Ԙ҂D{ NIqNj$h/{Yj$'gJ:vTu:yD&=1!="l9^W`\j۫־Y"[ҨEiHeȹ08z DO o /0FH!iѡin>V'!Wܘ^9rz%Meq{uwOՀLGYPs.u9/ˡ,}0:DžN?nD&Ɠcp|vT.$mu[T$'En\mУ9GMfdݨ2~\>-C[YWc⻂bBrbÖ~1]b9V]WImp\cluq8F%vP/JoC;o/&g_ gGhNQ.&e U_6r5 ›2Ί2xݘrnNܘI꺵-qqX6Wu ᧷m<-_061{~Mt_/ nfS8͏ ?*`^nVEXaO27RJ]aݣ ~#gM\fG{q?g\S~͗y_˛oȽNS!g Jj?I]I#{9_O^e 7?4_@>~y@insB8U|]޻쩤D to|a5BGt||]sa>ۿIʫ?љpߟЯ qYΝz$Yp_n8poޘ,C}9<3o (o):-.(~Et y5ZWa8Ep|UVXeϘYc{)(KJ_NI9q1Cp͎z՛᤮dɓi9+QLrY4xz8'hk=%"}!xE40KB$Sx $NY(Ff1j,wN.<?m&ѼAd|Hmy_ʫZ٠P(H-S!Xvr{׎Վ:9\K YEjO/m蠶nq2nY 3޶LFMLs<-VtU|tھޛb~s &19EQ  (abZd(8i1jN/c>*l4,bpq%%.nb]3t}3j}3}RFy' ,_:amlnuu+.*luMqFr_Gc_R/[_ Fr,?T 8'V{_gom?~π7׸G`=n ' ÇB}?Moߴu5͛fot Zg״D.-ھ_b,G'{= ?'>]}XS|ز5Q+gҕ0+44O:%Fũ\!9LaDqqM6DQOd > >9ν^` %k[}S&Yas:Si:Ó"(W&,"o@A:eRU H|le"HݹN=: mN<0HtA2sxw١Zw[LF8$AG?,p'υ)'y^)r-,x*'p™"QAs~0'X5,ڥ6l\ g+?xS H ъЙ41qش>Q= d!y,775U;|GD%lyp8p&2S#=87PǗ]9^?_B-^W "ױ{uo(&W\p|E;g#TOELie+e C\Ky 8 Kbv/~q#dEA "n_7]+tvM6joy1,?I^[nꗲa^4qѯޯG_UήVy5ګ$2~gz5N*{ KjT ^wrQ]Igczt5^sƌ>KBti#y|]o,->ŧqx퐤2^$Ive%$cz(=5&h6?&>xEW@f2r,Qm:Յ霏uMkaƃF&p` r'ZZnA&dCH1 ",Ǖ^Yg |N6UxzuwȵG7;$0Jb:@"~:\wzOQcic6ezHGds=Q$ssUTKN [+dqՉ% r!dRU_r425'2\0qo|Ea⫷ΆDsPNPQ J T Woŵ91o RP 2MN'l AԗW%-k^ utΙg(Iড়qD+TTHA)-X,.X4͎8(T8)3=/V,]2dN*Ԙ gZe/0 Ʉ=,{{kuYoJ}M[Q 7oy,ෝ~s 0Bx/*9:c)V[H2/4~P `I8..J[v(,+w F Q.ITʤ9cK!qh "Hⴲ)02U;HT4$u[GԂQ YìHm68[0UYwul͑ΰ_ЖGhg(Yh2kk܍[FCH\!` qU5mW-eMW%cz6 |pBókv⪻>o ԇW\q;+C_=HΠM*WTmWMWN\R)hzFʄлi%M墴O{oVlua2т^Y7?ƣiukY~ =bN9RP6r#j%tM0ѷc3./ūa8 R.Y|1_{a\on]q/ң| =+ܱl#]y[ZsS+xOɧ%ovT=T3{l QSЕ{лȧ_j pxG7>CCaUnrJWz"! 8ZBW NWKJWϑb |B"uG;@ɮCtr<< -T  w;eD+&JU7tp ]MrKW?pa'`/ OԾj/=?b+VzK]{ ۡ+ul&DNW@ɇJULϓ#N؟ya%pod/٫L]ګ1B9k;ki~?Ļ[mkՄx+t1tjYUvOf ?xD?] c$#J5G|ӳb0Ѵ]o.9:\m/>K)}_Lۋo5f-q](@(~go Ywu?>,,@E D;ݲ03p@}eycf{G /34.Gzy?Dl)|ُ[;/1`_uOJrfGײP|I_^:z>fc)0?xuor}@{Eo~͗c4u~q|?TI4|K(H,ζP6$Y$Ep=c+!ْ SδT )WLnyVB.ڜ6L>O>LRgH!~\hP|mC! L˜D% 1D&pC: %%=pN1#%ICKAh^01ZOfh!'+Վ W4j+D4!9ݿׯ}-4Z@@T1T݃6dC8dlWC|c,c16F 26{oMcjiJa:q ud?䋿wg7s%[šz T,JKT}wP9}z8o.B9*`o[ B9ӑJj6åh dk }n恶|X@{GҧL;α'94O,HI?w2KhJK!ɚ`/DmMs/*9u+)LEOɭE6-H} 9Y=<\hف1K(#mtAڳdhXhJZGvivt&kr@,+MV->1rM r3B,TTPtԡ-!xy9Հ/+(qa0P&PyUŷٕA[8![Fhcn 㒯e ,&NSqk ;Jc[`;R#wˮ C@pXk$ )ٰQ,jPѶ"jQB-ɶ7ePAqi4\ҴLobs oDX2qvLѭ]O J+9!_1r3ki e2a[̷A( iMB‚(!2iL~RTh>ç[Pbt,,xį:M;&q b`LΡ4*)ԙpCFh CA2K< pX +Ttgc(@Hq`Q ڳl;!JPA/u4k8傂4X'L\.qQv J*HO6EFAsyǒ`;CBpRE(Pή 6C :b@$Fu{0H6CZQ8b!z_4yYȠΜψHA;t7ho{ .M3i NHNoVU9`vND( /I~ Xy/˫~q<5T:UV11Qwu`H6#&f3x:p20ºd.q t56bv fU cJ"T$2urk't\5><8D;n0BI>zcW& 7[Fގ{qVwfj:Ncgc|XfR}Dq/zLqcc=fo! `inW^ͽx^K vG絾99٤]^^I/>=xp|۫ӫ/n2F6?/N]|I60G' |?m|˻o_wu89QL:݁ xv0C>;Ewm}{u-q ^Gҭ0cʳYGW'WZ2Zwp* :_% )ނl\3a%V|,CGBޯ7M[Ѱ ; ׹G9焝22[C }5ꉒ{; nQD(0"4@{so4D[ߜ,WdᾟT"d}_|)RN<5vu3=?&[Ϙ5PZxrD(-r`%Ƙs4*y?3N_f6?rhQ5mKW?8k͹G9Xo4i{:G@?5 ɞDKrt[_ o'^h#?wGAVXvS0Z_ m{|sw[5F_O{w&rrv'8!S%F/ YۘS\}Ts5|TsK!V ZWkEϓ< c43}ƪ3}Ro@/S&rF 3]h'Q/8|[NKv.NR֊fjm7 zBCj|BՔɕG*sVS^A'-PBGW\}4Ŵ2 T:zpυᘎА`CU&UVO:S)YWaꨌA$XvX*SkT)R WHIB%GW\z4L8xT.;z=p pzLp VOɅ PBjTR++C(P8"QߊL.*'8up\\5F.~ҫ}72+NMysr'it=fֵBDoc5~!KYȸHuRD F!yVtщ] V'Or^0ŢfAfX-À) Ero'2gv2{׋,xsz{;O)6?Vz4gm^+:.+:pYyE%XUTbF䊹 L]"ecOqsHmYmޙ_= _ٺ U{|Sϵtʫ~BNm1Sn.'7Cノ*K"wG4+ϗCI@TFO,R\2\zlH;稄rkt&]T@JT+( ȹ+\eeC+I$ŹpB\A^Ͷĵ!987gǿ{ݞdOn'ʏ>Zg/|h/'j>|72:_,_n;ڇaggH(8bVs=eYnڂ|i>pC(\΍Ho:-"dmv6~o ֲYoɾ% +n̊TμB &0sʑd*ᘯ?rFcI8!C(8VEA+h)ou:erDxE1ޥEA@hʠmS ԨPT?1ީdb@5:| "1THjjD9`[#Qu S1rvBY.ZMfFq}kz Tح'[Y7l-CC.־Dn$m2b}V]ڬm:&H)x+d ﮚ7s82J6z@;[wv3탷;tf~Fƛ]|.<n'ŋ[nk޲dS: s֜Mv뮧kl㩯nfҧbf|]cMe˭KjG\¹pٺet] H{r2Ċ1z% 4\Ȳ.9`-@%d^.({2eMM{ Pz$ Fm%4؞4 @8)g`ZΉ(@s9USܤp2)#Wm=ր]"MD*H')#)$YT0,sƒBnEl$ TK:)$m9l×8?fNCm7mA#R蘃-~D\ dm2,bj*@Eru8]Ƣ3(.I8rn*!7V>BCS#%RGlX-,ddw^x۫v픂B Πz2`"SeG.檅V>SQ G0Օ{S\=ZTC~_DBimܡIk8֨ؐmSL>$mj`ZY(󛚲eWZ.[v8JFqg)XKoBEXPp AZZ0;c"x'PHGu:pkMuVӔUcۋE\EszKq7e5>_aoe 3zujiwk޶T-}6? Gkr307 +x~ g|v𺹽 ;mtq@9%kY-iK$mg>aȥD5 e3zJ%B<Z)x%WkUT.8<*UB2QkjHHȀĵ@ J Q>=.8scc IJ9~T%3⩣|biô҄y~]@CxBopTٹa:Ca֍yWTgR1s4hμٱ* N 4pi+2"kl唐dԜN =o$oE1㈠28ڈRPs$4ZS6P@kF$>'ٮbRSe[WH6Kz۪Z ϋqO*dZ PݲmiGϤ\|/pwhLJҭTT7IP IiI*xpp=Q:AvQs꜋J+;+YT'DU1\xېPe1#ܕ`/*DL &Ⴉh,ZB 6SXs,FΎcg}u9}B\ss]-w]Ovf6SBWn=jzܳTp)o=]Mc][^.H俘.< Q izaotTi%KP(iLh4YTc\XFz P$3HRF[-YRz`օϯ|uM%>VM͏vEgY>ei(p9mh ;kD5AƓ*OAIT^*Q)y%r-+o ЖH mP yhPcP ueX%C ( ɦV9֣ ֎vڨ~y/vڹH7mj}8!TFy.p)ѡ't0dJR-I2B ц4E HGϒBSTh%T9|3Rl~\ZǷKB[b|g[eGip{;W.W95PyY:%$#[ ƔrE0n*(j⚂Q>/ tB9.@H2AbPsw-R؄ UC-bJ1,,{beP  ot^exMW5'_'=5Y*;> GlIyJ "4̡FIHo M9=Z]+AQu>Ҳ}9$5dcFDz  ]%A3tBێ Z0a0*P:Š9ۍa<.ڴCNV!'Њ2 "^,Q4^B (&=H*BF͐ =&dqc cQu!@<QgَQ?P싈0";DF,!FfE<tIL;VuH%U'"J*$hG)85Iӂ,eqyQEE5=ZDSx|١EĥvYCu}qQEb8- rD:ZIw CLAŴAѠF $+Ha˹qǾx(  vxW< Oi<@RQ8%T QHж!0)Lj Mw.7hc ;p!"IQ x%8zT RRZ.}Hԃ 9(2,Yt.T:Dc8);Ŝ(aNV8+dҳn[71)_ҷ%hvROSwVhqs7ڻ!y3 *v̧xSWxF#= |$]AhDAQ$*r\?\1ud6?;zv竳*]t1tI.D \ [bt hgH so$uݸq+v'9ǃz)k)fkeʹicn~1KƃM&;Jb|^#law: ǹL*kT.A:>x=9U~Y{x0W_?2>_|L'!Ek ,ɟ&0cS0fOm:U>ptT˚f/&`yF,}Q}'=\᫣_az U~5e'f4aB a'%͢z0YF'v26YL+ i|F~?n Kw@Q\woGK6"4`|cJ9Tn`Q`D\R"[÷,A *`$-N?BuN,$Ǐ* 庯 3ki>1#ֶ̚>*ӈaWtwB~+53OC0}C-@9E]2>?/FaJӿaaJPOƄ]Ƥ/whVyʹEkI$w Rsձ@M`Π_xSzY['}l>0Phl]0H@!y5+JNМn<` pqS{)@ypE AJՒ"@ѷ"SAMvp8-<__QOo=|gMJḁg1+:PA8F{Sވ-G*Fw0.eO%Don=z"J)cħ]{JA)@9H.`RݲnA65"G^(L &-DžzBE3K\ArOUrTnBy 1"wX?kdQ*:SnzRʮ2_L{soFYwmh[e^MuiFlXYVlx쵲J2a^!X *1j@{Ψjg$ޘZŨ+WZIݙ0d݄i>uM (IesE\9s6`Rzւ{D-Qkmkgr=0m:)Ͷ0`4#k7N!efcV@ (0/.%6!`BL1jveI0aAO~,8+ nEtA8RFdʗ DZC\6E.W iWՕ:R΢˕9%rPxHY@wڃiKdbʹr0ǝuHJxb| n/./E*#R"%1 s2JaQT]I%PD0@AP(@2qVX! nvf"辉u{TH=.@2x.أ9XzG!쵶^п*DӗM+eASPR%N3qGU&NM;F,F8WA0";Bꗐֺ> b@$yɵ<aUTYadrFʩqA=ٟO] 2-x CMV;Ce I"?ERPX|o2rLl?u޼k fEaB`㸄mTzp9oT #7|0aQ# cf:)MJ ]:8 IJ #7%8ŵZmMsBAV *U_oKsjZgu^ͳIhY9:"G'R}K@NTbP2F#1~Hg]Ðad0чbR(XIOvˋ1GgMSsTndר]sUɱZIs@Mf.u}^Y}Mj'|,uZ9TN;̍'xӣ'Ϟϗ?l!&jx u+0~K{.m&k?Њ54M[e^g, j;|8q[h Z)Dž2*] RޤN=a˅xĮ '6Y]uTf颂_WbD\@|BR&ҩp3U`.ke';oN24 >ZNJqFot/Yg!*w/phY^^Q_b2~[)s.*1*o|Hԇ,% Ii~ {w jnQJ+IMx@|Y:IÃwQh6ݶmE^,ˣVw_a%-JGZ(q&oOg}x%<Ҁq B>3c=IdW#ik$fE,g#]moIr+ ߻Z"ACv737'IJ{ޅ{DJ⋥DɃyiNfXSUE:zbLs1އDu[E@#x8$g2BGJGTh,E#sMjQswoMňVᨕ}8 iN^>\uTj-Jb:kɕjri`;z*ݝlҺ6Ay-_?msU"nS”x.>$aRbt6sI`$:T&ٳ Ţ:^|j}2I$4a씱00Ɵ?~jys&D 3i%/!isNN\KmrIY۴$RY+^'Yf䑁 ,Bl0HL9D- AKU9$ϱI'Gx-N*l"N!B(K l{[u*B.p6Ɣꭤzf$_MVRV[-hm;=-wDb=3Kjv|yNl*zd b %'SZ'y4]ƥkgb+=m74A=^tqmo!6-2H3zmz >?UnN nn棷OOs81͘~l' lں^yo|+#d423oUNE?17my} YX ˨Nd*!mk562C S: 1)Vmf"t /%nb˾'KLH 8d ' /[G+Sd fDu; 踤|]k>nc;2!`VGRLHdE(eIL\9 egcGdcǝn|!2i`[% %ԅ_(Hn˨ݷ#6T۰uEWwXz T&2Z'0T"!!2 f*MUgchƓ?>?ZOƓB>lC =k҃WϙV<;gkH&K% *<(|ƆE6CL u&g IsrLR!$:N9a(t0ɨr`X;8-AzySګ.mw[9ڇ}y}~EYg9.+jzCq~-{o"u]oo7B9p=sA~a~MvOB6A0F J : Ȓ_)v/-Ŀ9W(1JE3V`%K,(otwP?ɓ"r8'zhӖ1c 90epHX2)8'4O!CrncJ|dv>}ܔq<[~y|oOJcB~D@fCR]-SbbFegkFlfrwDtYyZɎr<2KC)J6.S Q)e CBւOxPdv;=9_;o%0'шArI!eGĽ@WMAPd/Y@0SR.aPN'/Ag<"X/ K8c#qLN 31qrxٰl;fGٻh5ujf5C!C*RT۳BJ|^ ?]hpfoƉyM Hɢm,;yOmĹ}Pˆd@ne_Jխ;w35߼̵P%ziYs% Eu{2/YziX>òGOE/EO_z=#ߧӋħi?,w$ F]N_pN.i|.KlN7P4=}}Yǿ\梖2O%*O+dϩ%} ~}]͛2o8)dO_ giW%e6\i>_߆)Es%a<4r|7k?%LMO.xC{F-O%=Y|Spf"{~˹?/l!kk5wɸ?<+**qVEc{9wz{?h1a)f Л^$7zCZʸ܊ZlBQe! [Q(VͶtm 52J)ga' ao[R+lh?+Y I[)Tr&#r)ʔR\&i:XG cBq^F<`|\jG'Rhy-]yF3-E/S:ެB0).^zV$>wQ+ c/jTJv ECb(+uH6rkwșWi@ϋC]tЅIuy2{ {Ih]5ֺly,ߌt@ a|U[M5[M3WU4@Uj6 LjgzyP7V7dNH^AzyXu_^fmvtqƎ`Ե® zm}?7 :^'l: 썷=vv#֤ 1DZ)TYN֪t!h2NF=39+ώC'=Am4֧A3,;'-۴?~~on2Kztt5yټa؇'x,+ vŚ9rsϵ֊ZZ+„hABEМoj/pESh5>$|oAt x $2BbtGpf, CYj Wc|V=Ҫyu< H^N=^^EkpM% _ֹ*h5  j#AH_Yhutn7[5B:D-nIZf}3O0gBȽaF1C B] e'pFGOD*zEL ѐܥą`Q /%&(4 #L!B(7rt0Ѧ[6#e4ԴʷA3rƷǭjЇ[@ns*K_=?FwM+)'J6P&r^lM4[&0 5]1u02woVR JBk96B7V1\D$G"ԗѶcP}'Iogl6-18MmA,#dt&ɍCҹxd#D`<ڒg覴oMO5-"H =J)h YPܺyfhβHJXAܑ8Cq +HKu =5s4I~JPѥ o0ya|dxk1{={ifyGQeN~18/vӫy3|樂uUߧ *F:qbf|1;~$h'C&,!Y =@@ ZGsvзQF{C80L9F&M{Ó;*nmtmyRs•Wđ[FepNxS]:pIWnϹVu''NMIW9PyfνI`>tg#;5CIЅ`;I ;]5[]| z*3fU!ΨBZ#ԶB4{u Օ6RWZHnQ]» l"*z'UU!XQW\mvE]j֣+R1WWP]Ya%c܆섺"r]QWZ]]*{uJ/LͫGŦȕFWq3qT*]J?B]齺Zs) J*j+Pk+ݫg:h;hESw.pTؒn6(CN^ߗeQ"9meY5xjCDXn|ڏdͫ6o O/..a'ttTMi,yA"3$^pwM9HIs,mR[#Ph$2K%NtY)e0J{An5N$6$t\ ›т9e8*Q\SEz'pSR#лPfUqese)=i !eɍeio2i|%fE[ 7sGpԏͶvAe/o}7 fW:8J0~Gk'q4.1 dVx f )"Ƞ?Iv֟H1M9?,w:>h#a( Xocj4IF<ɽ@3`Xd\zIK.*m~h{ΜYσ@GY,i32@2l.rFk{M`\[ 8`tܞT\ikRlOٙo& [=#cVY5mB62Dҿ)Mzr˫i!oopSJhDXHP1f!ȒZ3A&'h1I~ڜ&ƀ ci~y9<#i722ЇV.֣]_t{S?ώף;OpoSMs/R~~bR6&ϕEV^%8X&gU+kwI3bprLarr\ъAIGB +Kpo|YY_8ʳjM|=Ugئh9v޲ܐeiY[ Fnŧɛ4Om!ȃmja -H lr*>*t\2NSiRV~a zj~,q^n¯4pH/o?|jUseiD-tu;^O `#|R 99LJ7CP*w'{V$^?"ޞr1.h)y򵻧P܁'e얞Д̗ LrZy׶X:g$Uڰ\1Y>ƬN+*D;Ռ {bǘ9S0H ,EƔ3ePR]kȹ]3vU:ӅqƞPnu .W]mqx3Iʯ},yh/7O~ݠpx4_ {.x# !`E4蚈3 t%%8$Q:83rnׇQ7F1b<Y׈[N!ĜZzG$k8{V0$k S>d;ՈPBʣڬ`˨g|'Xhi):#v,R{ЋCuv%Eѱ^^6`SV< &(bڂh1kc˹׋^<}wePA65/EG |9}ݏ0̨4ɛX_+ahL:z ,J,fc `lt/5\:C`dI8OGyNT0OɓvWEN>Es A0Œfxcx5WnWt-4sAOd=4JY;'a Wu {==Kb`49Ł3~W~Y4~yU~Ihb4*q.'8xtW _#ilVSUw0Eb4Uy {iy(G|r=#u+œؕ!6iLf~h2;=>q֡0 F>ae=:x=zL6p]uonq37MG \Ű1)c9*;AFg~2;6^X),*Xul/b"L9{@N(yó75B@ ݳX$cpقXl 0"Hy 0B#X,0fƅJ$ɩMbcP>ȫݻvHՐr 773y1y|ЌK e]O?$0b-B5VjgxW@9l薫ne4-=ޯ721sgLBc®n1iWTm[t[oI@LRNth}KjŪ3hwަGΟlbE4j@K@3`GN5хjfnsr:޾rH 7<>*X!Q¤@)HaO{ IQif1\7Ey[!_9쪾˳+RSSx3'&Fx/m^ύ}e7J90bC7z%nqz}PI}gʸfr2b12ZVJPb;V4k)Ͷ$( a@ *ed}?YRv׌S6^t̥ I[Σ sMe,JR'na:XG X1I0&>_*,oV:_p~( T+EdrC`5RFTe5:_uF[:hP#/ީ~WKnmdri vmb/Np2Ojf ؟S"@|b'ϧ'uO~o^ Sߧn+r0? $'ߟnu. 8ݍJoIEsE"l;O˻2UcOI䟍_T /hFWvg$izVlo[;~I^M?gޙKo81ږE_9n5#8 õ.0A"d+HmwݧӉv:E=:ɡ8՚xẗLb=>dؘc!XTD$OXOP0mt $Rn+_lSM]J10q{c89IĐF|y9;V,gzsӵ)9|!ķ.^ʺ}4+.KƻTe֣MgNu2ay(Q0eB@PfM4ZaMPN| Zws*1%Ahô:kCXhJ5HDr Iq݉u}(ٖ94{=,L=i1.,&=7jLF( C==հ|Rx2_{$<0d^ E*cuyst#j@2PNєjaÑӅ_/ͳFpV/KNjXF6+EfȍW0!pR攧'EM29jv0?ez[9ԏ`*EoM&9:x}vjC.jRSjjK+{(UqЯީi=iZ~<\5_<_g ǨY=ܾ5 ZYo\q* BF# yaX0}ufy S9t "r|1GŶ3uTF|ɶQ.P:ɴXk2 W1=I}dNҩlT֏a4=v۫^|ϗN~Ջg߼<̜<`"HX[}Ine>Ͽ}h#[͍eNߺƵWnCDƭ QXq*xVƯG Wːφ`t. .1DFAhMS R!X2e? Q ]Iu4QvG;I $g @  ZGKsv 7cQ{NQAd1nBTyE%=st:rJEdJjs ^6!#x%u99xbϜ>jf.;}*Fxi>Hi( gНđ;-aGa,QI@JDb6BzlԒ+'Z"ÜL gm Zzye>6>7U InE.pY9n  h>iSFEL1Vz9o|+.uu)V!V7WYz_Y>]c}'*,-*q4\kJۖ^_rl/H&N- 7sxl8)ŷ ي֠XP-i6Z᧝R}&H({TKh@fn 0ZKyD`B8 <+H^X+|{}7>R ה \eC\G+Ti@ EPJbE~ ˥sǺ.묭͂,2e:5Ƅ̼?JN팜M7#KxSZk wf;LX!͛`}>>Vp8-=IUhPUQFz!r\vdNHGDE.D "3p'2u,uJZZ!:2DEpQ23"h%F{]LMuOPW ZJ懷ϋkLۺkz Dw *[!~jm+WY-AsQd  AHU*yU4 .2$ MI]֖= M޵ۿBY._0~ʼDz)YjJ(3aͰ:cLl*C"ǺmvF(@Duɚ̽::rJGZ̕ws_@ýUQHZJݥӛIX,"MmRxz-:8xZlf ILOW0[Ht[ LjDnUV UKۡuzݪٝnUUt!( :F1᜜,8 Rvu:*#A Ҹ(&*sj׭j%I]D .A+Q4&Ԃ~z]Bz9b@H.œqg ȃG )h2dZLj% T3{ѺP%!2/X?^AҎ2`$`2$f4|sm8 }ykNEgŁB.e6>[b<INEV .&i +/x; >KtrVV VBN/b^bLۘ{9 =Ӷ3xGٿhe=V2q!3ةd;W:IT8!#NP]<e2A OtLkt0,FZŌ@FL+|0aF8Vuy5ŖT+p8[/v1Uħ ͿYH7@1\4@ ,n!J8i2thuw^R >j_#7xy j0 R~m7\[^&Ny{x5cH,-{:]-cE0I;BN``'cCEH"(']N{`Q&"@(YD䘒. 9aw2-P18-W*qAL-Z\]Jg5Q[/jafN56g#3|YhGp&޼ooLU)c~'c?> zғ[hyi)R^l6ݟMr*Zӏ/8u*ѳéCf r;sKYDo|C^AV|lw"3~2KG]:.[?XlFd.SоMW={ODK*J> SvdJ"t<L"ƅ evYL' [ݜqqww᣷\xXl]oEYM/5eN 9"gZNe:+l{.\pԧ^Vs{*J}&MB)z9&'t8B2FJfM-uL6:Y` t].5Y)!2=N_`NRY5sn :Îds| ڠ(@JȭW)Z,:L\`EiVW,[ CN[4$i\:PN6!j^TPtZ)ݼԶh;PjHL;L"}5v'h3뤔Smu|d7P ZNދj@~u">컐WvZߵi﮽9LbۘS1-{-0t[ ѭ_) /3|` yfezc:t3?̖I'_2Φѐ\RmX\d88$/M-҄eO{Oa< ze׵?2Z!T1- S$I2T ZU֮dv;QP8\g,:ll9*9eYVL0C*UJ]A]!y֕%Ӑ<- ҐiZ_ \%-"~Xk7ze+oEVߊ[+oEVߊNwڌ4]Ϧt=glΫz6],M׳z6]Ϧt=glmcq76暮glM׳z6]Ϧt=gl-y VE|ZO"oU䭊U*VE(fEi Z5A&hUj"MЪ Z5OiVMl Z5A&hZFjV`;PQ Zt&hUjVMЪ1-0-n_tGE*Vݿv|6>}V1Ь3BN:!f c5Zߎ%FRlք*2L{3b̒Ыq2,E̙8eX͜;6U*հ̸'*^ kXU,">8?+xMf糰xkrthl[sŔ!PVxbiNe< u`l@Y·붧(P=Vؔ半 l\T/-IYC,^R˜ONsWP 49S(\DA81_!xo!_ sNQ'[3 5W!3䊢 n" $08,PP~'\͜+ 0 "VEDQECĆܘ!$s$hj-C^yDd8y@g|ȶEVl"( d @t4)TAPC5sc.ΖBۻYmUe\ .nbD@[!0U΀॥)Nn)GKA%F$,fXȹw͎⡮~x+W"1|l5g9AiCIp--A*|Cy| Bd/cU_>foNZ{KG~v2=1G:LfƭvB]28k#>wa"߮2Yi/9Eoa,+2)V~Ӂ+Ag {?,@_O 9zHs7ybc}xȫò:]TJc.`v8MObFg@<;p>3?=//!QlJh,88:MW-hA5~z-.{dɗ,m%L+h#]0IuJDٌ3!`VZs.V,KhV+zAtn?E2qN,0,'"{cK!Ƹ Q$/V伍"~Qڥ,{&C-h7h>Η6;~ӴOaл~x8{AzG&b<%c6gJj/ e./q O h2(#G,)c fGy)h8&Ѵrhw !}fM1-E'_f]hrX.wz/]c_8xe3#Z|H<w,0 EO4x,g2 zBOC qA!!q\P]RH*q/-8^`H ^/QZbx^zE/ (K8LA :fe2 ȳaWn*Xw5C!C]vd7H#+ ^x1KЉn6 8/ϼ$ zOje'r):ȁ}n0"V6~WeA:3iYo$c蘮}! "Q{24, tRNq;?`2?GN}pNO ِƧW4G8~,jK_޵q$e_v `wl b\bxH.I._̐dQc(Έc N]uUuuG?qbI2,m0SKoﳳL _Po_#Dݰ?_Su";s#Ϟ?]:qt9Z:EΆ/EY^\OU{YqG ޽Eȧ%>\)DK`YH82f`R!{jz~2\~}MɡGhuQӥ G7T0>Io~>'uE_)q٥OyB`>!%͢0/.hrx40_pbȦ! IOF^jM~1|<yw yAeG| h5{ B=h=OL&j,K]Takxc h1-f)fƸbJ0: LOm y#?;XA|Y?h?%P5X˜n0nvLbK7&xlGc;ǹ V6*jT*$XM1VBeR݃JwLiƙtdYO쎑$bLF[׵ڹDh"4W \߹{7SЭ]6bbNjbc4Q1A}0XxTQ͍Q:vF:~&i87\oLcеӴ?v&́L0A8AΕeNF],`Pix`,(3xn23Ńt= kb/Lih}߬ &gT3a:9 sB#F Q1JJXkMlJa^` ^ wE%K웮 ˜^ijl}>8ANJ7>^(L]fPU~L~?^4Wb\;0GH37D~ >|c̱|6PXs-9˝R'MvRԾ|JY C_nǫOڥA /޴MsNgB} o|{nto_ `䊜/0}iyC}e \|V/pq"Kȱt"ة>R}h$XA{ OkQU ;W2:E )np09sH[GIJieP64I+Z_ .&DZVQR113`NLj9+P8;ӎ.M7VIp-Oϓtgt=u)Nۖ\kw Ur&X^Alg :J{e)# ( s,GEk%/h~׌%Lb2SST--(@\yW]pjVz;pZo^KQ&@{{'&gIҘ}]OPǓ˓RWٗ^Kk7zA!@Og^J6bX9Ю͕>wxibl t+|{@AQS,-v5fr;Z}(j* ,`}iF\|7NN31̱]0#2HXG"\JlB Oc*\YJy>zlgL 6mmuǭNaI/NW_yh n14z%L+(L+' sYϨcR /:|2wgٞHFǤ"8n03,*͸Jb,aF! L|;6@%ü; z#sϝ ٚ(Vr=# ㍥{D^sDS)niziAL`mqdiVY'& #@쫠IqHN*D-@GS]GaBfIhm8@I%YB$hW !_3AFv>uiH;e~3{@$y]?y.,+f$d:?M70Ki{׮a;Ut*O]x`ȼgKnirH>TSDNB Ntt$xu: dEaB`xP~U 3tHߝWΓUe1*J0|,ll "z zD&V7%C q&Ͽ8'T$֊KXBF*$٭wnY~)*.|=>8[V+?eٷT풃DmXV^]sa!%%ojU3du3ru3&QQЃJ>>;zzf?{^*A:UzW`eҁix 9 >M\z+filTL~cB?P2|XX^\| ?9y/ߜczv z`\ \ai nv郯O7x]MC{b4][WF+zp$m-bQ>C?u)yb6 ߬vuHf3? i~UQ]TT@N<#++$G;I9r;hD\ 4hs;)1 oBT:Í GT990=nUJeS'& X:Ddd KL1NP87ݾN=:NZLlYz-S jSye᝗NAU]A" DET&0e- M4L(" 0P2P0%5tp#m+@NWR`ҕEt#\BWƟH(Jsd+Y{wT*Z;]%Ӯ]5^ Dk%LWHFh1=]m`,RЕjۡ~ͧ+,U5tp9m+@+;]JX::FRZ/گgA^ {65CoG;;*G UTh(=e(y"/΋iٍS$ax"1Kǩ=eR(ڙΉ0}/?V0#)6gF\s6;˜id 1b:)w#CH/Hh'奕u)»NyWcr>vA/;[ *QNGr²"qb޽=.fq2Q}eH01}$G`o{-`A[c\P[,[r %E%wa"R]-tEw<J:1`O#[o]t2Jh%i:]%K;]]3&J[atJ( JH)[]%\c*xUBٴ4]= ] )J0' JpJ74ZΚNW 贫c+n MSxk*Ŭ-tВƇ']#]i-XUc %v]5^ %3|hj37C{t lZ:]鎮T})z?WE\UMn{57PK'(-ETTnH#39#V&mX\5hLh2ޖ3gB)D~ Y:ݼ%8CtcV/YS5ZDX9 ț%4P9FY6̱j ]%DZNW %G]!]1p 7,ik ,BW -%MBBg lUBh*䴣#+(mҮ05t*Jhn:]JEt\c0@]%K \::H &c05tp%m4>RQ]=ha[o\r` $S!硯R(ٱgՐXsvk6 !UzޝhYLƝs qO|BZ9\-:řk7z9DrdR\f+=w )WM^H1{G,l y1}{ȶfsEiӅEզ-z}вeNx"ar6VwE|ҥ@Mrv,h`NB` Tr,F^{%J +9gq3-<)pȕFE@KW_.W;ʸ[Lr`yܕxBZk+\نM^FDgH\FW+wWJ6wurwHB`GWYRڰzQMP/xri+F])-Ov3ڙnIW`?ڕ4\) \@QCi&+^%W{U-J7z䊞8ti$s),S2SYF{eie -+CO.d=O#W,rZGvRJ6:G)ӵ'X p>OjI@&5m/٩eerGgKnV&  )s@+re|lOe2\wsrpcQƕ=w+ħ~"Kcɕ:EJ)es+cLr`f#Wn׼2&Wg(W^q3]R8%zM4\y)%mkW(WAĺ(0y y~rC6: 8GkYj-]r,*y&+觑+F;h_\ mrȕ}K㝸?k`pS"tphӉndXB&]re7zГdDrESֹ g+z.W 6:GJ-Os}⮓]du2Y'wkذE!)m!;K+U/>oYƸE- ^k=O^8ʤ+I:6=֧)_^`]XhrJVȝa!ũ84r)"WMu'29d&R4U=NI4\)6,[l|J&+G;U= fR.WLfwrCaN#WqWJW_ 1\\ac iRSֹvr"mrurKS]ʠ"WJWv7:CJ4ap9\-'z2ʵwȕCOɄcWs>{M2{߷x>˒y1N?7. V֞?i˟g?rXqR++Q|D􇷗<}TCmeߑkw/?]?u?wc{oow}tq[rE~{oN"+`w2MzQ\&(kO/JҖ^09$WJȕ8\)-˕R J T;O8287h k+84\yfbPq9"WJJ) *LrLJq4 h鏮\MPIɆJDyعvDk}77b(&ѴvwC}}W}PgW1M[4Ou^Rӽ_E~}9Ͻd7л[on![#~(F?۽{曋"X^K޳࿇W5%o^]~9 M.xg?ыj`>S>O }3pkͫBבl~{DV}xvRçR0/ $Q>1-:/>o>?'3S~f0} >C\AC51~w _ӟw?~߿VUI=[#윕8:Zj/3Sqc&_);Əb w|/,~!;/o{{?(gήTI4|K(LIYm!:l1IIp=c+!ْ SδT )WLnycJȥ4[ᅍx|}+CMb>̧;yA5}0 cn%0KyBpomΌi9E26?&m=)0JHɢ-An4ʣrC -6:XY^7?Ku(X?Lft '%=pN1#%IC֟KCh^01ZOfx!'+Վ W4jЂ}g"̜.=h2vXc4 hl.Ff:DmcR c0 cY%.1b59xK5~kb1BAud?S&+*-SX(/PTX -$!wu 鬪L/m>ALG*F CF|zs_,xL=dws=!v28Ǟ易2\-~F_[2dT&bCc_*T )%BmMs/*tߺL@KAX'䎚E6-H} 9Y=<\hA1K(# uAgаH! s'# L|YS8Pa;}b(1a9g0أY[@B^ 4s8?o]?8XNƅḀ7C/iW"Dmqql5KC;zT5PDžĕ.( nhJ}.+nF@p5ٰQ'`5h[5( d{zpz rG߅( ;O-FC3#\-" c" +ʄQtLX(FBťEn(pNZ25XQcrvL’0ѻlWrB;PcfX57]Ґ?!e v0J͵DŽu0Ȅ i03KRA[Pb ,,tu0wLL\CLΡ֤ L8r!JA4aA7 d&yX +Ttgc( HqaQ gwB"=dH_(H?k8傂T7P+_C0tsj B1kY5 2+=d] ӧ=hcfdYg}B Z4l=x_Q>6qC ha#:3 3aPourGwOЋqiIb68!99x>3$rTvND( /I>&#vЙE~{;='d_kW0o3?uWwA Fo3bja&a=Ìw .X\q:d.q t56bvڴLUcX}LC=GNa%¢. Ao%JC$JEV2q,{GhKa9`Q %(!t_5Rv.ݺ[aq2q z U]8pt5ڞwb:@vTVF]+#H b!c9?}{s_Ѽȓ,!Jq؂٧K6"ZwprIr Q LɣB- ā !u ,,6(}a-5KH)䌊a a<QyL PRC+p`YM[LrT( @0be@R5-3PEV]USgzI^gXyae C,TB @JGa0⓪W=kU'X` ̏!36,g1"7H`F{jRCU_85'˃LT LG0Քqe o  @ Mg Tf ǁYko~,gfphY E WυX4O aF[c|p]`!ީF8ØcR/,~\'Z8楀»Cf7>F P4pRt뒳O^}q,ll:,r_jܥ &)]M9,S'yNa.pd50 Yq6aΠ=;=ž0<8ͧ?pr5[tX?m`lQ3QWٽ<N- Tf0BgH<'TCW3g)TaH+PJ.ȀH DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@:^%I սQZ^ as:F%FHR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)W }R`#%ն7J @k+=B {'j6K ';1CnԜ;@ zپ .W2S5nLޫ.^ea\21 z~^^O .)|s=P+: ;c]Qjmjb5nMM \آncmSwuK\pk`Xo6WB9XLiyozz0懂/E*sxԀc'F9pُ7c7mE<$$y$#IIHG<$$y$#IIHG<$$y$#IIHG<$$y$#IIHG<$$y$#IIHG<$$y$;bIL>I<#FhK$;͹g^)R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)G Yi=g[ZoŻc,5^we~QPTJnu4/K>% .!\ћb.\Pz.^bk.U BY"Z}t(!:BRl]\t_ J{t(":B҂/0GWXȻBNWR;#+#7tp}oA@˟;1tBc++Gt ]!\ߛT;U:]!JI1ҕ)#B? VarB Q*򮎑Te p}+@kt֯̎CoNRap ?|ZB Caѕy]=W:#NU/th8tBO[<,c G8u>h8/'!W|7jDp c0¡߾ZQB< Ǔv6Lt^҇x_fEzX:NFeˏ4NGjR7Iox$qBc^Ki!XB0tN<W(FJi|ȵP>md >bFFYע|M6*1D}S'YJX4̦;mö2^)F"%YnժIΉ@?xިx#Pgڼ Qn>%vatjgLak7l~hAM,  @= 6RO~tãO&sZ)> 0XzDWX}+DxHtuz*]y,H+{Sh;5tICoOj/]= {ayCq/CWC)]=vU҈V ]!ZkDWHWiJpQfaju~?>x(Ҝ^/:wF3^.k؟&?y$vއZ%]e=\ς >Φ~\̺D\-4?v )nc46Ә6 Ua`upf舛 ë4! 9Y5 $Ij'.RA,,'=[.&qqٳn AqߟnxM7v^ bQ;'o~piylfwx +I5Wc.7 ?!Wݿ@$w~?mWq'_7y8ݘr.[ǯVM[4~]cU)[X0!\1r& iC/&p^^'BǓR,(S`\u9ƌ¥%]v.dsy C'H7 4Yi%)(x˒FJZcZcMb-3h#M,Ӓ&e ^1NFUO{v>ig:{q;U;6=EC<+*n!RƲ^Yzx=N *ֺ.057h``'hEmb8㾜hl`=KyVF|vq9*uv YLs0u|z[S3@Wۿl1=\Yb[zmmK+1w렜is=\Zcq*A6f{< ,x-F@k|݃h;dL11 s{r׌=* =)9vǟ 9jZJx*@ ΁iN i#|~> Jr3O9Jp.hTY6r0f5(سO784S ;(Ԑv_hSX:LOzܭ+r?*mI2kOo>mRZ-c6pTiۦ7_gk;. qE;d _,]4"fcpҸt 2إW\,=JSAL'Ow/w} sbEov<䞤~V/ po)yFcfQR2_C(z[X"5>hJ6ait'4*.cN!Fϕ*\\C5Q ?|Lh7>lFV%bU+2lXt6TM=ow굹.%|ža쌱/iFWMT )X<$S!dq{}L9jPd/*S3` Zsr5celaZi﫶g.aAm*=Ŕbf&u1qWCܖ("pq>bdBPu6D@Peu>J0]K)T0b$$hb0 9ӶH W>X$ͱc%GIPt;A|t=? byS`MCSH7~7 on{Aax?aJ̚q#"mE98ƹ6k "9"M$'b&͌C&fwWX/Di]f24;8sS1|(}} {[r FGlJ_ͯnL_wHp_Q&#S!cDYdOAf0mL[)]g/ƴJ{]Tf} \-.8=vI i0NҤ %,B ZAe/ϓwɖ2 w5,s+4X޾*1rozp@m< t?7um. IY )YW۞>1Y.#8!dU^Hqٌ?~?R17E٨txMDIERYYvqvq2י{B6e4*;JkN8w!]d7 Tu\ńڣt>EgMV(Q0;,B2kgW[d KX(87: ٽϳ؇!?@2i^=|$)=E0I>.Q*Kg');U@kڨiMoJEϩT+YgA0[O3}q{-[Lv`J;ݳ2m蘔I( ekRl A:ABIvY'HX7:)f`beIgY)&/#͒ %1c(XYQMfEM%&,%|) ŦɞŦ{SjޛV`P/ q|sdsn霹BYIb "zrvgZM4[#-g0j L)S@̒ l+-!hi>4tfs|&#U2V-rajdi k mAm2:]e-9.9x4_|[s)CFI;WYitg#Hq3t>`uKYc(dGdˊb6%]Q m.MZr&c -5-vN9]M:byo{K9S !I.=81_" .8 Vtc(}!2X4dĢn" l}C:2$Rզs=lIx1]c_(*[D["e^bLIx5d3Z` ɽFeIx0cp7NZIED.v&ĀI[taD>kGT-ZX$zՁ|]DKjR]Tbowv&S$R/Y 8=u4:"pbsoWv+tkue{0a+M?m/ >#V=bMs13 Ah$(u&F5&&`!ҠG²wNr^5Ml~< P,"s)`91mKC!C'24߾̵P%ҹG8.$Jv2d 6 Ӥ|Nq?W*o_ߧ_bӀt#%H;P{=ޟ'i1|.Ӡlޖ+͇*Ԥb97Prx7OK:8`.9<>o:om>g:zǣU뤮m\NG"r\hU$vqáL$LgGNH9NJGOhW=FOgŔ\SzNeJGycG&E3bCEّZ X1`m;̎:u//ff@`ujס5IW#&EWm!}kONOJ3!X;z?m/\q Bq]9ϩ3'xr)(P3ZQso>; o81ӵFWz޾x)֙[j}Cjd4&NyWg,2PBk),;u5e5<$h_ɭWNBN/v/o>SRw?dgt|?~ͽf+8l7K~ߒ 4rAHHKBg}a: ϖ=.)᳿x}/ F["Hવt: ] ]:~,eyF6q@ bkV} _~(]|:c[q}Z,6og|i|tpvJ+m9&Il*dgg%pdS'c5i{ǚ @iSZ5DD)&'[r{k{T{HϚgMޣ4y5sig`U܎rvmX%wم=s?9Ye[ 9B'P cdN 8eJzZߣVM =r|%yq :6>y2mqy$SFf$l+al4m|qxQud XN !>wFk~ug}lޙO0v-!gt2))ڦ6YxI:9B1swR]{ă^ƺ>*N:7asf8q4 @fw`?03{Lt˜92C XE;N^M& 'P7?nwWurLUG\!V'&31䭵c&EN; dȽ O_O_s#2[EGK?[AHpk· <zcC$)( di,8*I}S"jK^[U Fu&Iq$4zD߶}|!'-VFzRWjӹYY_M}>dN0.^3Gb¸Fw. ?fOI6 ilg >eWz6yglW[>e" G@,`(M6:&m,]zV:.B2SdYP|.+nF)4;"K k,vjӹp]B_@ϭW?;_u"E>}F r0:pZ~,zAJ83АII\hkBW*=O*MuLūOrbe[z >eRcDsQ>3*1N92=Om[r^ܸٳ^ߜu8V>xm'i2{WFJxeއ`_N 6,`l[,yt$ _U![;dGcU{8P~1!bc'I~ \p|{=˾mq0Gm X4瑤:BBJ'qaslxpMF#͠8o=5Ѫfl2dSV2~;ے#A'W>v^Um&4wA9=;d.N`wO^5`JcjĴUEnR_wt1Ej֩`МHuQ2e HqVvH6g2^SDȜRs :,X,:k`(p)`d1 V 1)׿5 s<"Ҁu(.fMiC2`, #q ?7<"0IC>dėP{'埃"撃ޛ GC^A403kDp0(WbRNXtNx/; y'm:{6Q$:{ :NÕp 'NP`)UЅ "Bu1^& Y+A>V=]FgIPoQuZoAoS+`)UH+ucx ~-T*.Nf GU0e;/Amm+/I ۏ炐U%% y{I竊aѫ ;PÂq$/&wUЕ2rwJ*uU]A^:H~,A|>yz:Meaޑ̬7('wq?;yu+ɿ߼}ջߎN޽~_&㪥U*}P6mp蟎~Yh#wU47/N;(ײ]VyE7G(ph{~1 7xb'[f9q BlPf8, TJ0DCA!lVYMNKc#]&6#IK%G+3!^`i; oƂ6:'Ș@ 8|ÜB{cG)9"*sԨ\82cDA_:0&d7ҵ:שQ;eٝƟk|ҋ㌯Uҝ,,C0痷av``+&H'.3AK,Yd=VKzc2dTxjVT vz4mnsYܯ0z;֤ړIe6q/94E~Vu-x:7/C㢝ۡl ELȰ .8 q-g3 p~aɀѨgiիw!lEҦc)Iqu I~{, 'z ].& ˯\oFaAn'h '(h='ql]`yc B՝-]=A`&+k m ]!Z*NWtJRPDWX BJ5%u+@i ҕ B4aM+,.ɨip@]S8;l nԀW;MdZl/ц(K\T̤.tJ >]tR+$B6 o ]!ZNWRt$paSv&g1$n4FH=?8Fxx1 A (.#5:L",Ʀ*gVgra~A%mJeٶբribkGLH~x~ngmQl&bW^kU2Y2Bfl`"Re# 4s`/2 ݦ lB"ËE#;vcվ>s^z}I>/{^[VgM-'mjOz^rgA!7}MΒ>.?a'aH+,|VТ̶P[qBx|o[K%=[=*wmVvܴlS܈BjwzVvwo>ݬMhsYJE}wJ!L+oToi+FQj@?|aN/֬|qKvLgHOUK|6/8qPwi/\UI<+٘2rQ+ 9/sZʴz;f\E`??[ ]jGӸxAœPUa\a@ή## ew=; 0:ó`Ū T'<5zXn0&$@ 3:w.,@ˬWzUFqcuұYPowS|$NpcEy`/a#W.Ra:V#E?Guwku6O޼]C-Pe:`tҔ/ߪֻwq7/~Z:.(/mZnѷk9_1qj@4#Cqɛhxe2~:]q~cv6~LJvQb`қ\v@1e sgMB@FgH2!vטc9%chWs%a7"V24Af8I>~jXt0Im CBbR&Se+QtMo3:?8yՏ8:54qnٝѬ5Kk{nu $V,(|xDs6ydKjfM{ -P- 87Y?VAۤH.fqc(DdXv*u4YԂ|fmPGQfT\f6*3{nninE vof,;ٻ6#W|;d~o!=}A֛BJ)KRzW=CҤ!%)Q|LqZ353UO?U]]2, bo0s@^aoy3BH1rW߼_jY smf^O[۟˶{,F>RlQoOlv}-^z n GK/S~8xqBoV-yAn_+yo-tkD[SɶcZ/Ĵ^~5 Y}MfӃo(ѳ`$"*U+ddQ,_>k愖MjH DkU:If<\@#&ƅ\ǘ$ Ij?1eLXό$! cL)@5qǫ,amSv]ʲaӾ pWA4}4wA.^W}&\V ׈~|6OPy?mt{ve?_l=)W-' r$6@ɛ] zyl}45?~{yפ5!eijo=_w~}| IPq]+Qʻxh8z\O|GW9_Z_9~svW5\M)xc**&:v`s 4zhrC~l0z`tn5AQ"r9;{wdZ{ueAIWDJ'c[,X_4%ss3T&fm3];Ȕىy?s]͇?@s$ \n{F}_5L?Rwgd6m&g!rYvp׆ F6AQ3b 8^EgI8T*""VOT\8_Y%hEOZϑ7N#eJPgzB a ӥ}ᅦY ;x\6$rZzJ׍{<,rlƳgXT- 1|~ Dsђh.AhXj:nkvq_"OL͆i-I r h,68Q[Kn%Tq ~DF~oi>g#˔o˃"̣Eyset9N 0opsv>"l^pQXGVe'1q;?cFi&zZ"Ge - ֧xЯĈ%]ߠ-<tDfYK?ϸ\,x/N^ 1%A A*~]FuX^g} Z,j_tZ{ն-]BmJ-2zj4n…DoztDRBf+cRę@ySejVu|_Ue:WR*ܾr]s!"*kY;}u^ Ia/cnYyL͞-B+N>Z hT' Qx=Qp>NHHET^Ġ[IӉ=AM'}} /'v?k369]/t1dqkR/*nSVDSelvxknicOp2Rث/zƮB(ۂ[γ|!ϒ_qh젮.bVɌ1JbB*T)Qy3,]H%'Prqj%P%Ky2ﻍ"ǏlJ+FFkM,ԁTos\ ([ 0i(Eh悷<M$&̢XR9* b<ҡ ٧TUTا-ɮQS\%oc\'E8K*$1CQ)ZD^3DJD- E ? ȟB4DQ&)E/fXJ݆/&R8ɥ$.O%㪘xoy׷.D6/;{gNRW1$*?&кQS)(W Z`mL)'YMV L6͊"*!}rhaq!1=k ( UDa$ۄ UqbXXL3Bepߎ>в=:3J::hI~ǩl~̀A:[ B*N.a) u@qXD8"s&+ȥ'47`&'Fm#uk| ĸYi (bo0ʳ@!_f8.ϕ8'X7_Cwpt;lawY 5iz)m/q>'=C7Lfx_o vxv \ǥJPFUvN? ޵6qcٿҥ/Ԥ%rz5o*Nv6𔹦HM9V|ZDI-mSV,6bEG$}Q&]A5&H~ E(3OKь^{3Y;ssg2\dI-5V{3.E1E*އ+ߴbc6RVl $2謏 } F9ᔕxn[W-l5ٛs+"㫅8R#&VԊ~|peck]\`avUGW,Yh#̣f?CU3BO)'@`g}X䕫ޞe|C:OX?7cX~DI\FZ\{v7|%L]º)Rxؘ`-:w58 yٿjlﰑ qAŃp}r5 ³Vูloy1,͕^l}-uXi>VM凳'?*y5Z4l2=o^M h_\0dt2<)f_}W\.@"r,~m)K7Ri%`_ c2:&t!Kڡ,9ϒss\9%c` xA_ɓp]A)X(4LjM]J10p{cp<7sTL 12#3Clۀ58[{XW B [ |;_zzM\7I16l:*2 F:䙳6lS&ʬS+ ʉ DkLL[1422 *PŶ= q*1%AhC NdchJ5`!$%-vzBD2%gVS./fA9FQ *M뿬kR2ne>^s'GKQ9B+4+]46~T?5O^M>x*]Fy0fl]튂至im-wF!Z[N656777:YBi49 FGwhx>jd0snl[wrS LE^<%L'>rWtfa^Q9KS`<.=L?G/_|%|蟯~u#ы8z9:F Bn2?lߴm5 ͛Flt Zf7{qHAX] |U_rlںQ,+Is'20F/Q I)Ā q;A`KQW֥=!aO,a!Y|B)Ҝ7cQ{NQAd1nBpS^Y?vQŶuu䔊0ZG9NQ Z6!#JJsS[O~۝gbӋbUoi8{LxC<"z#6  "r:9F-i,1tRdL8ٻ|>_BG|o{Wt˹]ߓ߇;mjr+EGVݛHH>8dJ7q1Z G}~Z:*1;JWxa{j|[oߦ;YD߹g;ZJt#3#rhU* ҃"(broCJ꛸qxKh","/訧2^_&8Q,ȂI (%-S4s"p_`~wI@xc\ w̸J|.K&(&Uhhyps2CRdh &F!>sd wRQ+!HhN4'gTF*UIk$^~b`m1>¢m˱D\BiQ p԰,}z@ DD4r6.&ſV?+^֪3R[dȾUrWgz#)c$^&|-R'KC)ErK_Nż>$|;gNit9EZj  9s He bHlUj 1'H## ,Qk…) 2T{*Xvdkpi'݊omS,~\և8;=2Ѭ4BRDKs,KX!X䴎lr'  0UzW̆%t2C_D9UU E}0Fٯn.~ůO 7QNxMX87@qZɪa (j1g2?% 1}l@<9Iqr,i ʝMך6jrtsutY웜轆aښۘi -s9o8=ؤU/)k˪W!=*LUߢ5WW~0F26[CJeSa8G5.\k|ͺB(YBr6[PϒlQJɫd<ӫ 4}f!ogTlt_o/JEYLyEy٠隆rC*>/sԵ`a 6f쀊U=Li{$MҜC;Lj:DkC`j:Ct ]!\BW֒] ҕҌ!Bv]+@(uBtJk3!BRtb]+DOք9 "\#BW]+@II|te zX+]`!hg ;j[zO0Uyt `NcՍR77B[5n[tEo@W۾zjCtભ-BFu$dFi{zte%_g+_rǟtIP8fIǣqn!wAXؕ~ & =Pv)@GlSŠ+GAAq}A~y^8NLtw~y=mIұ2gtaޔ.\}T2fW)qEn3ߨM*|=}l-8{͝kK?&jr7:rYYƖ*c|M6nѾ 5xwJVWL5WTuFRm/V70&l7Tzx> &ucArYEoYZ2ޥRQ桴ڤq_C{ſ-n_^x|E 4/^\w?G}_͸Cx㇃Wx,Q,~fc2HUqA8qpAW"֣GE2.8yg*>Z.E~ z"o\?}?}42q_LI$O [/F+M#|8MoІKq2:]mY+jZ3LcWXySiuz+7Rz:rc_K_5韸02QK-L2:r9UxTa*}4 < amZMXCϚ:ؾ+wuk_}~˗wMK,׿"?$ا C?wȶ$MSLӲFP#s$XKsԩalo/+! Xk _>~v%w/[:_ܞ ͻ)s/>sK} dwyDb|4W{]rVDvW/|&;}pzw:sW-+au R^'L?caMtd7 q~OEts<0?TZQWFՋ˷ׇ5%:||W=[V̎Z^_]¨xan{&026 @]yE)UL1Z@ E[߃9Y݃i6V)'ʯC1ruVmfEVULI`'|PN>7)zeepی{_9Mf_=?T֣G?tdo4BxZ넽"k A_aAz Y#Xْɀ}e(d+[)&R֣tF^12/6GL:9 e׊w'Nm'HԨ(2jAPa1PRڡ6Ca?CK;!DW*.K+AKqt%(9 %+o,G] \K+Ag c+gkZ] `+CWևӕ bʛiIJ{^ ] C0m=]y:HWJr<n1t%h79UDDWv9 \Z>w (݉~ ;޾Us<pV(Ơ F[{ς'Dѕ>[ ] C BiNtutUOpͷ9fUkc`Q)NZdS] پ'ҊLZ_sڻ^h.Btooj?+xJΙ:V =#MpE2[nھbJNuú +b pGэC{ơ$u#+o7+ӕ! Jӕ41"DW8,,pcX ]?rs;9DW߅[̂ ASѕW?t%(M<ҕgv.;^ ] ܰ'6I]%]*kbPbVss+4꤮3..jNWҝZ?]юʱvolܡj\KWZuǡotE#NtAѕ&Z ] \Еjt%(>1Uj}U Ru7|~7ՠGYom rrN2e`K<|:'}N \>~oĸ"T `6ǹWr2S%wr~AtbJbO6P6':BJ35n`bJ/-^] JNtut%L%ѕRJ:?w>OWT \kBWgДt|#+$*cj-n\Lh}JPS*Y] ?y\!̝e<HW1j5F&.bAAg߻J>DWԡ8ơ9]]kzr;w|{=EǸx@t-)蕥Oq)S;1 |y<(!<)+%X}J we ߆i\RhZоiZPM#M2Fr+uF/{;]e':FB;DWZ ] \ZLYjq(C8ҕSyI z1t%pMX ] Z7{DWGAWV!gÉo>]U`5gwtK1n ' m"vno{׿|NڻC]UY_wP/ӷWm5.Yz_ǒnSl_Q>mޠuP]]k8BltF`\~=%֐"fwJW\9pwÝ5ц>37!ǂuso{{= r!ќኟ-~]4o~Od;U͋TT /(2t~^717ῥ3=0C\@9}Jח ݴUmXV[l? [2z(Zá7킭u5YY⸻+:YK!xi=$?n^ nt+~gaPWj3\JRQwW#гn*T2ؒlHW2^5hrʤC5PbJUYlا)՚nqaԽ.}04ٚ Y'{?ОuV P67Q=q2}L9& 4J.rS&UXZ`6Fk۫N=hS|6n-6XHY>@R R̥fcYysd6CIqiJM#be|91f8 o\\\L5CёBE{))w@xƣ5"~uphQJ=@ܽ穲d,Tx2;LQ] BȥЈ*|/Da̎ {t Au"I۫TCVT)l1^ ,sbrX\>O͹;oBUE{r'gj*uϬs!+Au@r=꜌%x~fF0%cvSt-j I:5K)%ŒoDM0H/* (-TrX9${QU-yKBCpڸ\=D@ XēvN-Rէ 搌5S468SƢ"7% >{/R.A 0jA 1Ȏ ޞF }m.5뎼0Q#_f E UC`,ܢ kWnBh:ԡmÊ# (%ʭ2T|]vU7 bʳեcMpmVW6"גuR6%rduL\s \/ *AE#)Mٰ֞C*x@ %ٻV<$[ŨlJ=RscO c[tVw5TM`Xddb]AА 8p%8%X( r HM餻P? TBg3 d2imB AvE=ಮ ++ȸALAAX-H( `-@ كHc':uGŨ3|5gAx`7ia;*q b *LF( )Pu>u9*x,Sz#3ص~.NV!joLtfV )t0GRFiҙ%\x)J6NO৔Qzcy]]T «1IUDI)꒜32ErTrWoABb~)Srj]%"qPBY3%Y,k HvM4y V*Кe8M23qr-5{YKUUYdbLByUD!F!N 06dr]JW<~\ nTAD="߳e]w ڦB$ ƛAy@yY+=$]I,T*#d`SQd0X(hR(39ByjN$\Pd"UB5QAvΆZH`F{'&A P#H܎m }ßOd!T?Q y E*U#pۆjBZ!8Tx0X!χ~I囸 ȓ>COE,+tmi,{ ԥ)G̤uԆHTʗPw}C(#L϶sG]I (ʠv 0FO%8CܖSBEKٖh @k^WHPvyR a[vi5 PHS:fWc~,LJ35VEI4$?<(Bjw7965\F? ʰ*)eR]![%VK|p?X-IgD$Q55Ԁʬ$޺3JP)J],U{97i QHE}+fՒ%!60X;Xv;iVǨaZS1֚BԦRY3I: !r!peR?\s4d0η;erXV>0Օ7Ѧ)9ROC,ͧU 'Wהʕtoַ8|Ğ.ݬ-,I {u  5ES13F3 B Ĭ&7$XdI R"g}$PH! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! t$$Tm0s! ֊[be%I $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $>@K\sK#< Vj$1@VJkBI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BxI k}"`Ck-< d&I#$4" $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@ߗ+!}0.^\(R5u0@45K`az.s TժXɨDp{M O\m+\k-\x1ʕe#9ab.#}b]b%(W?\'6:!HA.Wl-W;ȁϳZ&#WYuJ W jߦ̊>U)yb%WZɻ.WJPQLYY6|8*I샙G?Z,ό׫q<,v e0[_Zݮ&t\OU*#9iOOz,juc`ѿw.rBnơt;xKmq9 Iݨv+ׯdNXU"W"[ymYYd^=}bu8tƓ|fGs .)ͫѲ,]}b[-ݴTiI9pVU\+\ n:gR% S-Qͧ,ĢǾnkڨ% ރ&V Nb&` s2ٲ9[6a+]xS}YS)%2h* 4ZD](#+*NQ =1rYG|yU/X+z&VL`&wr#kg ̕7rK \F\\ A St%?r2֛e\ȕ,Xi@:BRi=+0X*R*ֲ'J,:Jgz$WZLo䪘{3wUrUĢ+#VGr"z#W\*"WZN.WJrure%7O+`ٟb}bT{rF^)] .4Xv3CY;&Wz(W6=e**pjkU]+RE(WG$WƂ^0rq5Xb i>YH׸{w`aELYShhMUдJ0o*Rr*ARrk#bsx z3fO >%ܷ߯x{e &ՠ^rúef֞H,Q~_[|`/wF]lt߹ۖMN{Ҍx`tp~E9)~j۫a\U95S(+,be)sZ:ߌp't4hin;j˱&J~sTZOa1\ٍؾEaQ:b˪^Ȫ |phh6QBi ,_[:LTmbYJ':uYT*ABȒDVVGd/@?ܹ3+c(<*cqd4 7kv?Y~ܺR !wW1?,>toדEvQϗ׹K4B#4clywUNzN_Q-//-/qqTFt94:d"j7YRa+TҩNg $8k̭6nf}eK=;=d'e"Xf^r 6ZKHLa䠔8Љ:m))9-+WcT%/>֥>VC&2\^rbX'RH!(-#R'\:->֡>6Mar]Oך |2}|*sa ]ͻ{xg׾ew31K_GӘw{كYO^z|Ug4k'+]ie8?\mgmt>x]y>O]11wCi Gc ĩ}҂O[1 B6(JOuLy|¬b8vOuwQx8;_;y<,(`4)c@|9OW/0"IJ`pk5}6ƟǓ/_ Պ6Ę݆^W՛_r۽n߹sHyžRc rXѫhtc<.ކ}6Nr_r^JoMm 6 u}H@ל#8[)Dӣ-T""Wī*eyǗZ }cۡcj3WT1g2:k6YqR#HmʩJ QfBju×1Υ ܺITgQΡJ Sk yL+Mxf-?fsxLL]2aVTʶgڏuv#>p&>@!RC arHS7.Tכ ƝIK}k*Sc*Ğ=N}Xr=&%a)9/Uf+<'6@w[ `G_f`Z,R\Q[.ʠ}93%ў}H+lgD${D#87d!=K}xϛ y9vaft"A3 _~\ɯc^t2 n>e C\A h&U#Ֆ6pOYzh<(a SO(mSE&5}x[6Gb8j߇8 j\Dv2}yw| ǹL*\`'/t6|)*f {9V_"Žt\һw߿T|ߊ!Eh ,^Ɵ͌Ltɚt@|XCF]~D' n>ټG?ൄ,5egf8 r }@8KE/`" Oxzd\xf1.Rz7v<NFDDNS3sg:6 0NK% KpE A `"z%-cpzUX(lc 6)⢇oxv;;J\Z!IpSlLPuc@>lrdbta%t4/l>ٕSOޯ^Xm2R֍%}I5Jx9! 0[\A!Qr"Pu1̕P.g $B4ol'gx]Kьm)FrԔzC:6tP$Hi)"= R bJy?vts~'onVOxy& $q.3O2`]R9J n"`Gl_''j+oޫ~ű9N]mmG]Ar[۵i7 {j?`JEUSY4;bK|{Ku *YJwTȏ\[FƓe3~^Ax/]W.wndNwc׍YҗӞ>ke-Md & ¼&BP5; UTsc%b9NxQ>^$OwoLcеSݯRh� );O`(FI+*%0qJ#Q6{wAFmQU ʃ i5Z3Ou=)-0`4#k7N!eggc@"7"Q0!S&aJ;f2Rʃ$'?}vꍜ5m7? ZFd̗ԅs|wE.n"7/KJQrE3MgZԜN9it<,s4UV^QVN渳I ߁2[1t~[v{}Q,R")9IfQ* JJJsI$w? +C6'w3{A-밿sg$p&E1u္\`"zc%"gGa2oM_fkZ+ 5(q 8"4Iuh0b1 ʀ!y RQ}$ x88hLh,W g01ՠd2K`y=|UF!ۅ[Ӑ~wӳ*uÙGv1Iy\[pfY{'$sl|?ָtSrb\ ߳\.At-y e!sPq>%l; I"r0ARP|w`2\A98#7ڿJ`pVt! KXFY"_„ I,NlM"bvR;JP0<\fSI3K0OM{ `ܔJvֺ l_64 UW2 ^q[HJSA8מp,5*FĻ[)]pr?9'|ɋޜ`Nvt u30 6.$02uZ񾺆ڮbt먛 __MrGGboeboҔ¿.)^aծ V0)] '@6Ee-'i﷩bD\Q.|!)sf@O邀8ҙ~617{Idv IF#fG˘I5hP}Bnf=;Z+}]l;~_o+=ŘybL*1*NF%<@rFϵ,wJÃ&O/ ZZ7߃=emd~g4&y [cR;ĬW)Gjd9HysQ  cwqr3;TE|{73#_fBྴ~6a+ջ^wy1Pt/e>%F='=Z(g ա:-FoAJ*#T6R>dh6*Y O9-r͙A6ƒEx!^zŮ(|{cGclq)&: rTDT څl6ZBD9DŽD6!źp+:2,"ubrX+?B"N9wnWPRsH+ޤLS{Lw*CXiB7X'ʛ9IyR?U]@* SijUϟ-ϟo^-|Av55ʅ~(rx7f`.HL)*Q٧͗^rіQhU$cU^Re,:u~i)np09sH[GI5leP{~w2:7A0ɵ r B($<Z՞2#,RF,=wrWN ۽'}԰i/AR :Lw'Lu;2}$TQD^6r#" 6mE].HK2lcؒ#ɓLWՒ%m xZ-&aUUU-O$>7cS׉~7;'TUQQyb)GS- B/ NIeQYEu 2 /5Ę"ß! %[e/\5xWCr& ޅ1 W8-XFoR 喼 _E֩՟N>ޭi.-"/#[5 ?Df=p}f̶A_=4>SJJo@dV2 ]>* :HCr21ӟzOrB]+lbلWBkL2>f!Cm1 Z  "œg{PA)^;Q6b s0fL(1BD{I6 r08]dH0v%kȂ`DR L QI{nr,kL5&kьa‚obyw ~Yq&2ejmX gƃ+:ɫIN A?C7ӏ6T ҷ/$L~c34vgqsiXfn,g W?~ pϥLVCeW BḴDX,sBŎ8c'=Oh{q>N9H*ՀD&E|:  ,KT79dSB,AP kJ%.PSf.s>wU t08(<茋&0RWiB !Ǭ!Bu4,q9=8]O~0vz2I=fmd\7c\t(yބ gY}@Q2]s9-NMELWoo|Ld XQRRbc079U6gx۰407kY[mGHDsj>Ge.C8(I]P `~nT7(OKXFǔ=ZG"1t0uőEoAjJCsj٤ΟfZuc{V*w2qN*㟵(0O,fl[6h' o6o><!;b|fˡE&^y|z| etKg{ mNa.zuɬ4P&cdmϙwWxwZ޽y|#yj8~MxgwVٺ^'-on~U:=Znݹ{>̦}Ʀy$}lh}G^.to>_4t;΢]nOעN_/:Z%[9~v$O@A0笘diWRZ }OF(!Y$Kɢr;"`("RBWV>#4b+ciT+Ѯe"C*S+k)eAtEeы+ɜMhm+B |+tI+ ccBBTzS+Ǿ "NCWcP ]ZaNWR/`ϡ3;wd}WЊ#T=]A:0 BԕC=AnhUʖ?`S+gb|E.Y3죹0}}x/jy1tYY:*IW9~9W.6#!It|5V8B5 ~_UZY;ަE{t~֔uSS;ӬgV:IPXEV5:9~ ;ƕ=\ FYY% T\Qc-O"Tu Ajft?bVM QmX/u:Q 5j\09 ?Wt5]DX}GW\Ť@XBO C;]!pO)lC 1䤓Ɣ&Gb誁 tE([:RV 6tE(ap"]ig ?]\J+B+{]J5)ҕeL[Y]!`]\[!Dk;]ʖ:0*IVr] ]Z3SRN*DWتbB1+G^mP >CWnϡwg9n3 OWygZs$U7gƠ@Wn=U]!`0"K+Bl Q::EsǙlr7gb6׉Ƿۏzki'gǵ-HHڋ*gN#T>\Cܚ\s`Z'U{v\Xg E)&͵ + .RAez7JaAEWXW ]!\#m)th-WP-@W/BWJpJ`uBX']"]i$+^ҞV ]\J+Do J9)ҕFAlD1tEp*-4 tuteӬ$cr\Wj'tE(aS+h\>~n7”BWVj+'7 BB@1tEpU1 PAbJ^3&GO 0#t+^2G(u]i?]-)CϥS7֌CWR w":Eυ uFLǎ ]H]~A.mduX VT:,>ٶ()M]wƅIu鷢 KBGb#5;AKN*Ŕ(5] ]!ZXA98Nꔢ$"BGhE+B9Oz]CW)tq> ҕ[]!`P"c:{cPNܖjG}+q#Z{o ʞSe 5$"A u+Diٰ3xt \DWؖc \\_]=1#A_XqvlW{'BYV)^7Jtԡj +l)wDuPJ tuBtl)+}'~1DT ) .ANQ5Cs eb+@(7 в$ 6P ]\u J:ERV\"CW@)tEhA9Q1Hm9Nd BY^u/Qpt,f׮e@WCW$ ".. hWVp$ @sS1P:6 ҕPؕs^zP1;]J9_]=^1%wGn:U=1 ]t+1S;[}i"Mԙ{K!Y7wj#\4}ia)gtr{au. $`kIkRGj "Ktd Cb*DuYÞ_!*{|i ~5Q~WDKb6j>m}3˿O7wk5yGyՃsOm^eR?fff(j3^ 6J]&J*U& XOF~3>r3<9+dL`1 &F|1Mh1(UߢcEiP m }3X0SgהۛHѤOeZ5xzW/տ^~#|O_4WnDZphۡUr@f.UB7~.[޴k?7i~?D>s )/\6?#Y46e|A'n-)' j^WOl3zҘ5x{7_MC--vL=WӋs&$iB/G6A:m%YSqZJc`|בS`h ") ~LGCe~do͍vKvu3j]9x4_ 'n*ѢrWo'TT6V_Ŏ?Onҳs.|v>^ ɸƞF)>_KYK:=r/li>Z7f.Y\auh-:cV  J"cpcMY_;:TV (dą`QRPdWC@ (r08dH ^ L,lkGnd\ce'lbZiIAeޚxm*ލ:tv7cA3S'浅PsȾ&on%9 )2`ى}֑[~TrLUIRp(.5Ե*a8HIQY&J9*!ָi癶Ir*/둌R=}f}-kK r:_|7pjߨ˻ΪKԟ6FfNnnfl2Bs@[w߱bj| J_-hi]f)^yKKr:}7j:yvx1MŇ5 zj{eV} ].7s$I.ÿY%ϱpz95[C!m5sK#e:6l!-F{t5ʇ,1 /3܁5˔B3 8ݚzCQ?>=lPn<-J81ivKrN/ޞn=\x l7]"lVM˝I'[o |:ym`K}[AfC8dKkBP?gUs02֪΀$|Uݦ^6}O)}͵wmmIWzp[ubbcמqv20PU h\jH,Ht;;ɬSUCrиl´pv~E7_iCC&Ql6I yԮPwq`p[ޤ<]~n~Oލ7_E-[v-tΧG|Cί{[ޙJ_sg)L|zv!wБzaC7S~<( "3^7_g_A^+ڠA3t-l{H 5sk18&fU^HF%]N@dLAD fQ422xHxNr%|(bcZV;_xEzyı{7EWRԻ\L'#q'\ DZ LY\X!΄;ɓC#Ӂs;WB?WrD'"3Cdc([e b>5c?I7/˕^I'tQu!6z}]ͮgic, Qle`1iB6R&ڄx ҁTY,ցKB:LV-8PҪ*j/X˜g} t2a=dXtߑW#eȕǚqqиQ&@&ۋ_m9ҲaJF `1 ۏPJv TPJ}: AO(4))N \~rBB+4%Q"2m ջR BU:ZdZ)i926%ނ xf1ZA.Ws]V3 N>!n SoO[G]rp`I(aڡCF3@Q*a %33)56$Q o19 $ %%h7a=fX˜^Oۡ"iصˏ^?U5?9PqY(:5O_L]PRΊKO j[\sR ֮d[?i5Q65@M.3hťȣs)"|YG haEUdh2'YJU#,d"&1Z8%]&̹fոJ5,bXz,<^YTm(vrgga5j/4n4 gOrO, Y`-d=. |H "@%=ștY^`SMZFm4r;FhMRf`jd7RSAjq* =j v(&Ofh4)%T$kx %qh<0o%xBϪx)\@RL$eA$CXKđs(Udd\1j6/L1`<D""Gwܘ!ĝ s(Q1U5=7Ze ![S-cDH(7NJ Yբ<I' F$ 2iR)ُjO8kVuJj^r(.ʸ(z\qq" yKeTb)xa)1娢RhqIV9sqqռPa,-DpԏV jV?ֿxU櫥H /~La2 bje/{hOr' kS "P6kDS2{ABQ&/ R?(+yWSQ(Zh;3Njo:Ͼ޷6}*eO5׳ /D{Ĭj)kCl c1I@M,pD(|sz`x8``MbE"3e"=IJD"F `d:2.3k%r 6-\Qp w%S g'ڛ\V3+O {J/g- .~;Ӕhﶈ6 츩e/|jOLS*2Cu?,O)D2 B2`V- 8OL籇N˅z#PSAH@ ,RJ๳E T<0*~zshGF'@PYY"Ƌ `6t2oB!kMբCa`N-:bq(dhQw"E3ozY/F4Ixz6oMVzgqܨm 0IyO0EH Sa0~{[*iq#;GǛRɾd"e1b^Qp!rEy~E1x2eK};y= n?NޝuzY~_>iڞgϖ&ӳby>-_~`%F F B0;r*/TPc"i">W֌b/^ ~7 g5O5R8xD8y/pt6z1m+}5y[3G%8o;?^vϥ*}ۗV-7Ng/ pq(rl1-m2Nф|UAEU*B. v/Fp+,[Kp|^4sY\SvJlesj\ٺR7Nƭ/gjj?&nwL11bǀ. _s"R ʎ=vIy4+Vԯj$?vW]!ܩ|B}Ķ9lq,f1gG3#x~dV&h67}vj696ǻӵ#jxdVZ:Ъ~3d&.xWg,]l&F]R{BTuweB@1F壯ZY=uϙ]]oRSFyrSB?=c9(?4~xN}vb")* U|sڋJ()x"*$BA΂q8Dˮb^ u3Kũ,*Q7uyc^FTVR&~xCͤ^Hl ;#/bQ\FliPch$RtN~vqfW+yڳꃞؑf9dKBaH Jen y.I35$l@n"u4PM9&~[*+q8aՓ8;hRke C+"m=5}r&BO`kckVϨ^5t? [rWsWb Ѕ,׬_ Mgdkkx\@J֝3c{?M|P˧a>Q8(YH66UdY\z-N(nR{>f JJH!j$1 O-csÓ@OI/6$ wpf-vg% =!qI0:bIܿ3o|d1 "D6KgnrvӴNDڋ.z=r/(;HXQr=&#EFКŲ֜dNj{kIU>ΞS1ol@- 6$UbR.@ F=@\ &3Xf d.HremIqg&Nnɪ⯪UZd6ps$9g)҄%,3D ;sF7Fmp8[YS*clÂ`;5-}YZGwyCEW.JddP@`0d(W$-*NO Ozz"hչ !&%CP@KdQ"¸&gRZZaN nAioh ` -RV7;˄wޕvNRnX*5#N `b$P{ H+ #00: L*p´59OB'myDb΢=z+l*|-DŽYL:V肉*SWYR J帲{AN<9+w,qkt0Үt0W㼱o Mʪ~.m JDE$t&[ 6VJܛ#8c!pK/MMi}On& @ِ"s4[IRc!A!X p Dc.&ʂV([(4LjM]1\h8AK;4 <2*,Ȕms(6 ÷FΚt\ԾóÆeFw΃%<LЎ cNRH )Bl NUAxCpV/:^Gڶltьm *2%OAhô:iCX!5:/4:rD-GvcPȸӄ}w3 u_Q 28&=!ȩA(DsA`'cz鹶K/9iYuh"Dx*T"R"RRD}CC* RQĤG|0p EISg!Br Sx4>?&yR(ݏϿ{sg>ւhz}] LjOteOY8(8I?'|͍0/]q}_WsV)ͳ%3 CC@+*|(.=T3ƭԠ129\38N~8pqArzoQ$: D"JĜ\ (Jj<'L:߇r *Ow!uq*E2%׳MWi%7gHQ4@>k.g_54)7 FŏWWG53Oc.ٽTwV7\MΧ byp_zsbn$j: ۏׂdsOB颩܍nfYޣ B +3g1Az6E+#w:dSMcEz8'3Kd_}5F >NcR<[\b .BO߿!|?O)3u|u307&pS~yߵmuͺFuNZZ|$Q~+{k%@d_ 8׳WekUˠ+0{F1#L_QQmS T* 5cYosM\9|޴60~4L*N'*4c1kj9 ٜXP i6Z᧵Rm RI grػdf0Zɉ`UɌpėyV? Ʉ:!K6 ޤGED?zdgC:'F(3 \%H*NF]j>ybizq.Xu&AU ^o&9/nWڏp)iZ0+HB9#RY%"3F*f$`TU:&Vf9BBr$ҔS)%' mjM0ԉnUf㉃O]xaUxAeJ̕s]~׍/;d>! Y#>_"+mhP2ÕAz53:mQ+ -Ds\ħd"Sc[]}TNƆMavn@7X;f;A'۳i#E$[ BiBAQZ: i"G|qt[ ';bBdQ`P(3ɟG@*q#N1&q#N%3xC-P/( )CV'PYֳ?FN Ec <>)tʕi'HBi)D{]"h.`g.g.dW|/R=' $ܞ=t7q\s (6NǾ%TT|l &Ѭ4BRAUM@L磨X=&I1pb8(FfE<5#}r1íRzC;A)fgn>s7C^޸GeHW WͨtdX=E( G2%p ΋{]Mnt;a{puuYR: Jy*#Oo>n&S [ HٺWa ZKw7 t%r=RWH0g]MoIr+i)?"2" 6 a}^l#b(Q&[#jqgTv- Ѭ~⽬Հe)t5в;u(d^]l]Ҫ Հy)t (ŕ^ ]1RGgCWKN@5HWSX ] u5ІRֵHWBQW;Z2'ԾCHWJ-- b pi9fhNW@)HW u9Kn\Z{'@h=2l).Y ] BWmS^^]EXe "f{Q2n= ~h5rD7V<ȱ:Hb7}PaSܵkqDS”lz0A_af{?)boN !xc%Mwm& Ӧ-:J;oo.@|~ir&8n~?D#)'i9"B.KLJ 7vzau}%j0d=#tANEbܨt'윜J:|@Lp-NW\N`QCZU]H d‚ ,:](/VŜ:]euK FyI8Հijxt5P!Օ^]}bV`rYpb@'>Pʪ^"]%] !,\Y ] u+УPFWztEǎL˩py1Z9NJWK+~+72Xs3QhI]2_╮;VINw EWnX ] ztCˤLy?dYNyRHgM<XJE4#h8Y^rDU‚tYpi1ax(qnxGZ?b\t5>wyqhEO+]@#т"CRje:u(Ůt銣0gqpл=*/9媮^$]w`}ະhɯ] +]@[_֮?:v)tttLR_]!vy +sPQ:9_/./GDK=i Ҷ=w͛۠ܟ.;]i?vG9y3W~G%c0ÿunwǏtyyU6UcލvǷoѸt^]#X1G5pwd9҇T.&>=u۷,Pfg @˸UqkY{֗LənB3lOKܔ<3K(2!I0kZcC qȎўRKͶ#oG0L#LˌBH~ 2H`9dgb{uPQAm> Z T9|xCߦh %Jr XC]S ±&6V7>b"$l00ur?sE^M G@POkXـlXkٌ`UhWEԎ]S:j:U'ԓ+e , R*snTP"p1JoWHJ-*@F.*  lBClHqN +𑪅\lJ᷒2+$_Wx,8KH&`WH*zbh0wB!ՠ໴\?@ 6N[VP p,Lh"lg󪥻 ߚe A'b΂G¬ ݄c46KQ0I)q$88)R  4XƩ#3m~.aV!7U-@Q"Ł.0͑& fY2RX6[goaHy޹ENR{FgU]%ѷԐ#zFyif|6L)H LJDFBy7'Y,] 4|a6PZᏍvmM4IB>Søz?r=/wnz1#.M5quHNo8*20B9}. vЙxaS.?lf,kn҇Qe??{*>ԂY/{.0A6}ɨ`&a-A /M>o:@Gֱ*GIWR]:\e ,#<rLeYBgE{jN$\`2*!Y 4;j 32]K1zOy&Ȓݗ]dUW[_F#H܎m ɬJrA#;ϋwl9 ld-Z!ATD4X!%=./m o"Ofb2&Xvo,B&uHc .y Ŭ>&H ^%1kPG}_I0SaʠvpΚ%8Cܖ()v%3cEX (|Pt ˈ ) ;hȳjFNH˽;Jfmqt3%VO$ZX4k =lC$#p54 k*Z[ުi"If -0y/^i^Y*qak@MWfz=z!@ DO5g`vLZpi䋞`ɍBfw4CS3Nz|dNzO^+8 NrlCɵ+FA;#@.`U9\<*w-j1ZrAvF@jr< 8 i`IWKl$dipyn'nqu#B`j0BA1 v^|uoo5US&,mV e'w˝jnװd ̛w_~yݡ^؛S^R7؝]vvIcG_[~{uپI>߱kba 0 _6z|>ٿ_lq퇛ׯW_3{L @ *(3.UŬ)q߽D7^ ^-TOA-Y`q8K|WCY⳴_Kiqjxԫc"/]%tj=iَv$I wpW զSOF.Sܢ*{(*Kkž+RQ)+-:<^ ]/ZawZhg1 T(ҝB츋^8p`Hc,m LMӎN?*(BTAH[^wb=1eqM^L+-pGVR8JbUfbb`BTV/+XTL&ȴQj+5ڢƶx*L+U)ޖ˯:mf2MZ=jb$VKQc6wQ4!kI|z[G%iLn,0٧m}30 c%^'?Wg5$vA'&ߑ%e% q ޼Xh'`vrLZBԡ䲴N.Kq'wr\,wW 0x q(*K]e)FwݕPRIq@ ք҃AWYZ.]e)EwݕTKsWR,8zYGWYJ讔㮲R*K'讴VR֙y0*塸,{wT2ڪJIPJ:9welU >Iwe U@`M`CqWYZ,`设wWz}LtG'5ZK\kwbG {fP4M4; wk+]);wzURtWO]Y!yGW.pFNс>N!2A3z x4ceU MR*jT*}#,,1Z„4;3jrÉ**'4.YTn.}swS{57?s܉9,8+s @Kd,CbpM-[Un\b}'jv5)zxY~U\[w%ߋ7+<7](m0Q(aK`5yWǽ,&_a_w#~w|~m/*Uw7=%~9HL/3X0Ȍ%ϗ{78/NLˍ'1,:G"BV|GZ( ȒEs"'@[b-LӀ f؈¤?:a3f Uf7wS3Y!h4I~&rEu(UiV|S]څi= 7͟:Tj^yxy[4g^wi,X!@> 0J('_FP0]qrrmQ$9U Iޥ\YjFW &fO0 p>9RyV4Yq/ڨ)4Y^;-Vٻa jN)?XPduj?eiNeZL n-mܦVy0gVOZ5{f1n}+{Γ<%v|^O[{W郦gAm2,Sq<M;Ǎ4ᣱ&BOZ0)xk"&DVzD9oR<@r}ݞWi!6X/d>~vpz(,{1((n~shf7s[۲NAWG\gtc3gpazFx'g\/!=X-.rQ ]9Lvާ?/~ x@L h3QK =hWZJJJ2xtېj+:\ Kׅqe K/A+׫@P^ۥoV[l3PLR ה_1X:ũҀ| )gRRyǺs|וTU,HG=ƘP15eZi+r5qB{1(ڨbix9ͼ=%\Ra/A94U5(.5D#A:aLG&l$KQ?LVpQQMДoURcse"41Hs!jVH A4&꡹`5ɪzetGᄌ>pJFRTj9ך8G&~y}oBZϭT…RqyKh}=, *9CZǫ`1[~lW҉:ZZ+]cjQ A{t6Ie+.J*UL*rLnd-W+E1)64B`5A)DrNOI'˘fIXJӜ \|yQfOl# u֗N5X߇,紻-{sWKʋZB湓8ɹ2l~;۞*H"GL,W?FmrǮ~}vvN=9ʲi%c%:&!ᰳb nlBkkʹrr\3xPPDϔ&ȥmF7%m`N8[Y˶>py_,FY &:$]7Q&H/~_O0σѪ:'"" XCqʮCM2]iqiWrrWY FVA`ͶA3 #cTOIgd'L[#hmED@|R/ab|Ȧg[6YԫxL.D򼲌% NZPE>XT烀dh i_5 :aK =5_: o)'| 'L@"gL0dG&&8K6$9:ܾ㶩y[5dNc Qp\ז:-zWOvcE@RWku8hRr *u[n_U֫cnw&ClEshθ?˼7ZqG[5});9:9GC{! Ud{<B3u Hr}mQEG7zTkKK'~*_+whb#"-v:aEh 0 &ؖ#u߯--ؔ8A2J7]b=|X/&o ꊽ6<}Gu%h7_)@24l<_з_GO,1?V/' Dc@sup2ݩMtL V‹~:Qy(x} O5W੊]cj=oG3wj~ua,KQ_FF/9Ng4*G?  Cuț|'?8Cg+mv(4[w uqx>NZeiWx'R4/X/zɌP,Yˤvf]vWŹ0gh=@|.?e^BSݦˡMc->%_VteYu!'oCG3r#~PzNg|y2,i+ބy*~589.DAm ??ѾT= 1!e֩.: 2v 0NN2:\vJ۟cl;m0'2M.}w~ϩmTMWrmWo=I\0~1g.X|k?| ]4O8XI QKrI@tӳWRWlH>:4J<]\b(8[斧| T:ɚ(gaKjz:_agҵżYp>_䣊+wG$_y+ʿ+FQ՝|zՏv|beytǣ篣a.w$m\B[ {D!. lelQKFPR)bdc#92ؖM[5&m?vG蕅Zrԗȴi=JhuU6}x6܀RaQ%a˲$t .E*4vۡRn'2Lw-\By$\HF#S1B )mOfR)r>'QX#)X fGD,XJf>c!.e?bOz7ͻvvav'1_l&|12 hXwK Dj3xgjLB`+ "ŋ@ %єଏ"02K 2jh(INhdx636 ßD* Vxm5Jp2_|[A::TV301 -T|9^QC"$H-'sE'>8}햔֫]lu,)F_0im$ mRB$lv^hy^:C:/=cUd˅\n{/jRmfY٧+U~1]GvLT|ò%Iq|Yt)?~(@*jRNjHJ?ң1VoXlpRg:4STjDSaeN;[;‡Q *GR邹(/*YPhjG= /5fĹOn|NfH=^_ΖB|V$7tO1WJ >%|V|(7V4Ͽ\=ǵer[Ր;󮸓OA)6Q֪C yeuhы<۷dPmhvǠDem d>)䏒T|ɵ Y%2DChwF}G q6f6X;MIv>vcL<P*cHJcB*YDuުbA9ɓXvEgʱ! pMSրbօP?eRX"B0A`zϰ8w+j~?뷊&+PIVVTl .Rk3ZHt)1-8J;LsWPԆGiƃ@RL-ɘA1SA<6)6MH訲eIUdJ$c!Dd:[zO8waoלAw͏"lrD7ƹ bYzLJi-m0罴F3I LH6bnɢ,gB 0lL1%]-M6>1[Mh&=Z:|SM.:͒"6E5∋S- Z18-r g\NFZFcF$<w͎n0<]4KZ"kobE7F?>O# Ƈ?c!V_rRYvU E7{oO~iWq lq>~|W NgQJaVv]yW E/%w{=5q.A ~X4p!ݦˡ׏f>%_Vje^zlXsOoId̻hW7XP;|}/o 3_hRkOx=WBԓqy6YR4jh==;<_%hG&h2v(vLY*4&~1bk@uk\n;'=}L:>L2[V l-KTt-y%(Ԑ$m$9.E?0ja7r9`d9A>^M]֡ߊS/ 7kك .@TL$C%cmMJx4Z<Bm<"V4ag!fXh4Ր1k.jBa)yUb`521՜"j)eQkK.,Rr*iwl`2d>]L{T )!gM qWҳ93nU=j=v͗ھ^">-oᡎ%F6>]RCej[G؃鎂X*oQULW#P( t^u!_GꓬE_;}8`sPEǎΫҁBǿNeLi.r*K0wtu.Nd1_zßl$F޴uJ~ww냾K`y].^쿖ߕ?Vs']_ίJ?/)__е|=rqo5?O-Ë́†P1ֶC[Fu-_x5~AɟL}q'qsgu1{}ss̾z{hcU)_+ NڗC.M=Oob>U)|Y?ķG}{?$6ǫ]?&9^wwfr/ög8esS)|?_𛂱f/.V7Whg%}w/vh|twMf q@7I6n]^a]X:a5mW,_o2^Bijܰn};J1>V^}u[.6c$gIw|LѩSqW NzM`?O P/6(B1L<|xbu3V;qf{ϑ~> ܙuU~\^u MPLI~ӇGA1;5.4L> ߞanMZ/aϼcN 9HU v|@>zf݋1~VZ&qCEotoްqa&3yfЙ'w{wSvx_mOK6Ґ"BOJWJe gu~vUbEeR& 7WKTpAu&: `k*tQ[ MpU**iǑuX7l*qsC&*;@|(N&?SL+-%g{om]esWT(ua#ʷ]]^]}ΦAjN 66.m@cʵ6buI+O2:e݄Rm*qym3/a߬D\/.kumMSy]yF;4۪ɍS)rXA໪-!s{kqXQ3'4~Kw>ux'+|77M7կGTC4%\ꫠ],ʶ튮;!@z R5?`k UMry&Xf"7']5} CS="CsWW~C?=5w;]bC:r{XUˣ-[m{t+jD{㬩}Sa]M"tN :}0OPc~maF ar{X+)tm=f֖qXxUК †Hnhv3\ JA,CcY8,ԕ`:mk#жe|KC붪M Ft96P5]dNvjBS>famAiH:*TQAWP7ȵtAy:˺Q.3vN ~h6BO憿;vs5NU *v"\RiBD]1չ]yТ ]1p+ EW^y>u]1YWUPR ZN)Q}tŴ.RLMukic+ڈ+u.u]1YWtp+ SZȺz6{v}[M._hZgsr](6פ9/\,W۶x^6/՘-՘uVy뭵-.Nr޶|tO.jx h؋F]È5F>`P o5r,:06 B^G7~򧯿p]k+hvUmu^FΞ{BzTTa [/l)ֆG,{=~ruU-uZԶt}4wV*3啛n/x3}u?XNsefj/6ktN(T[3eMpQE|ۇ[ &k%&c\R"9=4#9t1Gr䌶J9Ab`gqyrO<1% j`xb`btEFY)bZSSu5A]9,F+' qJb+Q<c ]1pD1b\+`SS45E]c4wAN0ȸh$ 2Σ)*Քv À V2Ȕ&+4 Z+5HZ)ult{v=)Q?w5 ؏w5w乫aьap0Юt'2 c+ܮu5]E k ^|B^tr ./٧.Ҡc$:+M4 ~Qk4n``oŌ770m 1&8n07 +N7)bZ(rYWԕuF'n4LNG)"Z u]1- j:r΃tEq5HӚg^Sԕw1zIsW3 =>rRtE&uE۬ *x (gp}RtEuUg]MPWSK]v1uE]m (1-YWO+銁Ӧ͔\};St㨣nԍӚ1j)銩YWt=g,#Z)buŔ+T%]hhְ&Re0`7Ň\ihhn?=+NޮE;YJ9AwMTIi+wutSaa]o(^J$GvM.SM13Q9te"(A AKp(˺,%͓Ȼ ƍbDuŔf]MPWV:+5!`(ɣ)cFR9 (EWDI]WL &jJ(IWǭ'57j'ZTɏ|uLA"`r))bZL^WzYWpF+Ɵl!az`pjKLW0@Wuuho}T]VwaSSzu5E]R`+p ^|Ԃ̦Ɲ\A'tssd\RLkcOI?VݟILz~x fw-/ٿ\ N i5Z60)Κ-E 銀U23$Z ɏ*@u4V銀qbtŸJ2buJR0֋:'EWL2d]MQWA}X]1p3wŸ(=ўuŔZe]MPWQђFW $@2.*)"ZuK&`Q׊y2ȴ>|m >JL?>غG^;6dpOm5+uuh׃V[AbpRRtŴZ+L-!itj;%zc{Mڌs L^Bx!a_)ua+;r Dh_QՐ ]@U8\w-l@AZ1(%cR9c(HW+#EWL&gJ&+Vt+6OPXb\-fhr)Ϻ^Ab`qFjI+VNz+'!0%$)*8p IW q}+=ui)&NFZoRtŴ.y]1sWS:vr{\+5O3%g+gכ3O`AAfka>_]CBk銁#xPJVuŔ&d]MQWI=-*E'I@J>B5j^f[4n 0aH70mP2I { ]1WbtF)bژzʬ) IWF,btŸIӺgeY) o׊uŔmɛu$ւtA7ѕO+u5A]ୗ rWT;Ӧ?bʘ3 ]0[pGA|-B"Jyt5E]a@a^N$ EWLDd]=]=ޞyA0`G08r086t5rkAttezީ(HW ]1yWhǪ~7o=ȺP)'Ԡ|ċ=[%^&kHς};JCz\^'h7 ^om҂~H s (E; hwDeH1.z)$?Ĕ[r$7H 銀qaJӎ>y %f]MPW\?@i9;qhmH]WL鲮+>hI8*+btŴSוAe]MPW1X'HW  +ƍb殈6*H]W=e]MQWQie$8 WN'b QYW*+wͪўuŔ!Lu#j箆њSFrt岮zUYA"`FW[)bZ )κ"+[[ 9;-z,'A1OƵQiJSbN;ӭ'0ÿcL[cv./ٿ݆75p0Az)fZk)C-DH `]1 RtŴ>Y !jr*A"`N)bS+t>j`[5+5b }4FGx0 W#zMtŔd]MPW($銀+ubtŴ1Ht*?Ң8A"\/&dZ4** $tA`D?o.brb-& ֧G^́_I_ι]QӾ;gV-ޭP_GmKg1t*e]^բ}߈R-5\,wo{9{wiwmi*/ٻ8#+ KK`aaU"DevKMET/5K žŤ޷Մ;~*j70SvG\f#G|a|!q#^߼]X>>5 '_薙y8p7_5?f`^ 6CACor9?叼. Ȼ>Ŷs56_W8RRφ5[k\]h[v}I*]un:[[1>%?~_A2x,]Eo^;${~^LKٛuU*[G MA5-DtV?b77^班dJ.c4Pj 2θKi6kT'R;ƈa4Bf@9wFBqxoݓ},MV1\΢z@֩ X2|,Rg|B9 A:^yr}4xNuuhaST{Hdt=·zs|XvIkNֹ[ʱ'Ws̉e g16c@FUymBiLHnIʗ VCJ 6Y ЗI bu÷4Вj4 I{m͋j#r}HF 9yR 37g ,(R>Y*b=) ;*D{ZGviE餍](P!D`%R(- -֍fR|WZ}^qDg() %JJo*|' Ţˋucm̭ĺDdT já<l 34~ :jojLEw%RI9n,BE{Ty'D)J6NZA!N ƺCu Rq*RQv/ J*H{m>#FAyǒ5N_*}AY8(Yl,@{C\4l=x_Q>VqC hA}paP/x  R5xlPBr|\1fPTԃ,0F!NHe׀9Xoؙxagv~1ߞ)^ׂGU0o3?8;5͈ G 1=*4*KP˦tdUm jmlAhqU +0똆'z$; .' ) >@(E&rZ#d^S1ˈoLÙ.̓1ZxxLPBb2$kuk+w<o 1t5¢JrC#QhzBb9¶l^ A; >$[y[ө pvw=G\d9o x@^,4zV6 `_[#;Z,Fvݚk:jFq5R 2j31rrp id <%2`O~@rXr* P.(7fhokqPDD;Lw[zP낂0dj*g':RC}Cշlb1 V$5sJmhB $\a~G"o Q0 bp\ 1\U -Nu#guK[sO*68&գ-19mp@୭EѼR+5֢Y u(P5jI:Lj='е-9V#/G&&K/Zi.!@6(=<VAPkMyQ6.4(u!u~Z)AHgk|T= z66PCX6:a$ɀn e[EJ+e2 ՠL>O{qv xTMS*%L2ńըm#7r fSOl~BP!}v l{q痒rf[ h&xG?{yݾ\6P`@6zaqn#N^&OZHý{ l>8Nntغy_ ]ݿcoS氇K[ ;y.d/:}BH"&'Zϭ\.ƸhY3(>JڅoyZ@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N dXz@/ !p'ГtYb'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q=]'Hqkr''@@II@O ĻR8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@O \HjMN pYpZhN N@O QDH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 @mZ6~9筦~ssՆ@kmw~S0@+Zq MZq pm\q h}%FK?qrs w;ߝצKWwC2EWt兮;:uZ]+Kz-tht(ztR} Oxܭ: f[~8-/ #uYpl2s=G?=sNnZw/_z}=/~ВwoUxoilܡ]]S|9Gye gٽ>}]-,\US1__ iZ[&O)1.V Qxwe]FvnB9t-ejS ɮ'=-w۳~뗦ˡ+DKڿ|P$z{1|ĖkUJB42%x5/Gƚ߯6m>u _ĹRrE{N6XʿY3r'Ξ(?_?l.ˏ6;nzjca=W+ ﴦTr W۵Trք^1J{zEtڮ.ZzQYxzt7Y]QjEz.1 n(ztسaVDW 8Yŵ5QZQWO| ++Z.-kZGOW"]a{L?p\MY*;]1Jc ]EdSVDWoWCW ׬f>=2ʕHW)ya~=FZR BW?][}xRn{1x7)|_=yPGv@WACcJ77moA]|C7-HJG%۸6tܯe3}'2%N1J"4'v5hϟ2?{ƍ ]]F*מIή\q*JgjpH9~k _8˲=@7׍P2b]:Z=YbtI⺨~jVy*lƸ;-o aq.ɖv{Hf 5#8~'Ǫ4ǯKЍG?C`lԶs=nL7?~uQ̎Z > YYhg^%\:JYI [ߦa9P`^WhlxWuFP,|r}`Wp:)1X^feީor\ VhWKgm5pV+YH*_mVj}?K3/żdKx4fW*?x㦗JNoijtd}1D¼wc R8EhmT53fjxX}ݦ m+S߬3TZ>,lW_JIx)R⅔Y?9 H'RМ: f9iAy4 YUObT#F'U8B_xYA*/8>8X m,W_gZw+'LSK r"h9GGs $UQBT§mqB=o#{brn @*I[@Rpd 0o3~2֎rtՊ[`MQx_|Xq~O7x/+nVGDs\s s;̓.ЗFd?C.\k.*51po:W6pYς2"񜙌P]kturdjx3(Ik<˻f%[$oukŹ۪co zOrr;ԫbSWZ>\[Ex$yO(D3#L>.`B?H3jWX=<]O9bSҿ׏33Xr& 4%g-Izyb8 HJ!Ыׯ=zz#~_.O}ӫk?k]{8 LN]t;/h!_A-;AyB:tTSjƻM"YYn-K{5%::i*9'«J3N{ǫ(%&58%p"rJWȘZ$!H7@}LFa7F q~8m/ZmxJS~ȉ @ibb29fpe!&ChR^X2R`zED{Z(cW&r[B,cEa)^ Ns* QF2'ٞ{ǷQ R^u  ) '˕W>š#:~H5\}~uҎV+JqQy[6EF'oN2$Zz(w(yw͞@MU1!ZM Qr{Q/Orv ഓɘdVIBJ QH8ZVPMhmmV$3e EF[ KYfb?j$ Hoڠ8{QjČnhn!K.+Dpto@F$HVC&prARkz>EKva~bO4Qj+s:Ĥd b)L"JDCXUfZgjttvBx?,ӗ{zW7sFJR0QJŹ֛;˄+,Uήlt_m ~,<aG6m:Sމ:b޶ |?EM VtAm k8wH%8# 0m>?6DEzc},|]8)U'ˈP«[^Zer\Yν ZQ<+HGot0үtɜ*+F~%|",RE2pcOؓH䈓y@1䉂M 1yTWbg,DQu! )1\Dz @鍾G @QkiPO"Yp[IRc!A!XB[UѨQ8NTYpX %R/*U>piԚ)c8p{whxeTY)҄Qi_܏JWr.,?FiǴft>Qƒ2z,H!-LOkp⠜2@g{dU16Yo(,LkS0-NPD2V`F# %":rD-Wq*5CX|0p:F%OopJil4O(1DN >HE$P k=GURM^r)"BQD:,U>AEDD*L(H >>?)"dix>?DISg!Br Sx4>?&^m|` OChrE~6Ab`fe>=t0PSD\IZߏw}.v!Y=T__ojKG& CC|pu4S0J('䚜ӊWUʏ>TS\NFrV /'N&n&W&zM.|"ʋE(:_"3r>~\z!iG Gm}Jtv<>~stEO$T ׿kR2n2PW\ %9GZ҃xvk7͛We*8@0|9WaFwr ZlZҕ %ݒ[ju7Ye(qFQ̚|\>Z9NuV꼓]vLEU>62>ߎ'uYWdO)=qs:+UvK yMeòy=Kdǿ߼7|㫋߼\o_@i .Vp_~x;MvoȾMs#vi^Cv$jnmBo/JzX1>8w=fRk6cq% ` b~,9UT}{*ԒpbC1 NlHm'Gar;Y+D  Z)9و1Іs)* ppw'N.}k `.OgB$\>N#p.p&$`IWRN?:O|!tnu:lUН ݱJH{N]iΐ'³P|PDPT3K%I 'x8Җk*IP8> 8-#E Tfõ5<2% '!:Ř#XLɔ6Fryo:QJߞ1boGd(6,(OO'AW>zb9.`3n;6dWٽCt7 Bʡ9|vV_PZP=trժλCI+wurG6GPfT.9YxhVHA;4%W5:D~tg3~\A;,M"4üP OU9'sYz?אMYo'\ibrH`%ymfH܊`Ωv8e-BYfܥߔD6:" |lMVߌͦkN{ܳ[lw 6$>=}thva>&'i]~jou.jrK%7W\vzE;t~*dW"V[akv٬^uAq+'luKMp=.Z}hbK-^SY&wyv~[꛿_yy7rYo>FMϐb[92zy5Mfﶀ~vuU?46NrikWً<ImbO @Ķ6P=gϚ3pgWz;K8C|i"?F~i@a2Fx3{j;(5ep,2GPf9 sΥggW yAv\mж﫻lcz5|L^v_gGMASYMi|oMN'f:ܖ,ǃ߿4c&6'4z= sbkٿzJ{3U뻞vt$0guA_-FxxqqE˟tA1I*JQȂҙ˷WPE,S4q )K1 ӽxpB:uᅎQ3D:htԚ(I(,!,?vpIP&P̷U Aă Nt9(SG ˛Ep= #{cCBt6C79#<>q㤜n|ZL~I / Q£0V1wrݽ< Sp)Z -`V$9s #RY%"3F*fj^1MΉU@'0K-hs7əHS"Zm&.1Mo܏ipM:nq(o .;FfoZ7/[A,anw _}]_H4ZJQf21HofeK+ -Dsmۗ h÷5Eq :-(Uߍo7pջ}euZOHv׍S#1p@ j4ZEs#' pTgZP* !>j%0e8hNd[W͸a]*Y37%,RRʈ\VD {o70kĹG$ \}[w0M=eYh@çar5?u>\Ldmo >r@z&ecM o{wda%Ⱦ)aįꒇ^4vo-1kX5(B$6s7[oH Qx'\p+CڜKᢻcpڔ!࣢*2T!HI}mBPJ (?R^H' \>rE1J;4%$(vߒ7zPvw/o #ߴS,꛹ISrp`I(aFJgD6)/,A J٬Pw'If,gV4% 6$QgSL$%)B#A{1 CeE&͊KՏfFni]ևlɞ //VV|ݽ{*fEvӐȣs)g뙌 LZ"Q>yϰzFYmDKs&E,epJLFe&-b*la-PJ6ƒ+RoI:ylUƍ3 DTH&5%e'/ Y`-d=P u>$RȒ=9nˋtZFm4T kMRf`%Ĺb$^}FG_6 V{`wx`4KQ gL'ŗ^" 9[HġSUa&sIX"1'BL;Q cL17^khE>, M-9f#qRR͚}@t`DIISC ĹEDړ]7ri(.vQ vq;K1)yK#?&Kb)xaicQE͍%h#2rbgb}l~&켈k?];G޸$2 QX7- 'dޔ|C ?Rby}m~g줥ɋg_e8CK靈X jJ2X5o+:~RKݴ h%Ld.#VZw8|̾gG+~?u$@:ІӁ59ɵ&mЮUϪϻOx:wN3Uk:kw\GWC_Nhuw]`gG+;LY,g F/_}3Z՗3WhT򍏨D@F4Gٹ٨s5xM0Isqm)יUh-Dl$ms) 3A_*G9S"Yaӯ<@A}N:62 Y$2] 90"XGH^0 )X!A'd~O"ssgugoOiga~^CoWƦOeS-x%ATzG:Ɯ6mtY;FߙJ DJ@`bX`1nx6 RxJj e# FG}N+q0VnZ},opl ҟ'NnV\nZA-v_+E":t7xOd40"34Z'x  aXDH̊Pɔ/ |g:[l5=Xoj0 (,Bʖxl!'lp<{bF2C*UN Z⮻0.Z|b8?xZo ,`2( 㴉{Иj1--,YlokEgop3tPwrj#S԰/0EF^N4ixzUۼ5AeJ;ˌF(]h#^`M'3EH S>Ljju[2i};;G&TrZwx& Yף-KV)$ 5bt&s#WvѻU };=> n.1pΝ|/˜4i?W/Wyiu茀|ub:;:./Ft_ߧ5(X cᡑƄ1FLNH=aŠ,%c"LNi}SVћ4^fzϴkjqFk?NY rg ߖS_;iCq}ZWlrLaq4wXa()f/n~Rn By2_mNVT- w-IvãWJ=#5-?nknmŊ)xcp![Zi\;Y;vtQݴ>Tfϯ~-MntK7b0(8.Dk6;ʁ] oi#N孭ŧBOm*5soӳ&sNtD-)(.pgwf99],Sy3'7"t=!QU/l2WI`g"P"uQ뀫-Եyxu~1joT6_ \Q=vwh ة si!QccXwe̙O)h^KMRSF)ڧR xkj<%Eȇ: 1(V~'g%K()x"*$BA΂q8DK ^1/PFYq3OY_JTO=x] gJ& ˤCIP"el&1lR惑gBtl׬Q6۲eNi5DW&Z~GW+[t^=v(dm,dѫTbMs'ʹ 8CI;sL"F5~%a[IOpsAR{\YarH>Qh4rE]y]꼺@ElV7Nݍ[HU0+J6zt HՠDV)q)&' ΫR=R#CBU;*k*'~*d`hn{/ĺ^z+{Ο5q>bkh{Qʣ\b\F˵1[ @Oyäp4G%E0Ǜ-=AS5w9y$CSV! '~ݲGvqRƅYe/<$ 4A5^Xd ډ~;9R{wPgK%vY띙V ^Z#+b` )xkfE 2dZy01qWsU٩?0E`J}RE\J)Pmo8J?ތYpd f VK/u;KnJkZqeq18CzXEr? ){ז@s Umq]1-e+uEWѕ銁I׊9hi]bJ_t5E]y ]1pbtq RtŴOSSt5A] ` ]1btq2RtŴsSg+c'?r#* J5#%iy']=AV">+ƕ+5&w]1%ꢫ)*(turjg3Vk$?⍺qu5['}Ng뫓n}Sֱ|uSYXcuuyqู~_]) h {*Ll ݼU DlkP>㄂W0 XYn8[8 y0663}6B?,-ЪplֈnP:J>mnm%#˪ՍШUaU{ %U/.x'=&?͇W:ir44Gtm=-j k%-Cl5kZZ% ya8ۃg928uc_yW FiH쯳'zoj3p7R|xć:ЕYUȡ2T*>f6 tf7(ݧ_nvy1@ ^=AԪ "Lp]/HW l]1RtŴsU$劮&+4H+ `q+%]WLij"r]1.i)ԯ)&+ڊ3qILtŴ.{]EJul]1.jgZP;S&7]MGWJ"1rtŸ(&bڧ+,cWU k*=+KW]1>dJ*UرÁ" '[Wi4r2FkGJ(]f`HU(zl9A"0 ᢑ+}]݆)]MQW> YƸxeАCT>ꥯMVm..Z'洼!!D[$*ZPIy1qj1i>) J0XÖ:I]1RtŴf+DF"Urw%i)m(ze$bbtqah]WLۆEW{ѕu LZ1bZ^WLYI9$]10yEϸ$FWLOҗWSԕpA"rƮ׈jgڱQ2r )t֊:1bڐ}t)A7FWvzP+5U0L5H;DJuu}mc6 UFbtŸ{}:HWL;DDJ늮?*{#peՉkҚS($,+$Ŵz֔ T믦dvT P #u_J=A\cOJ^"39P2 fr;ؒ]13RtŴ+ Xt5A]H8'ļcZ^WLj+vr+(EW?b^FPt5]X{9cW]EڱK2 ar#iA$ Vx S/R6HvubtŴ!7ҫ2|t;V=(** [/FjO#%iyR V=(kFw;fU#mHfUEW{UP1Z _n|@KB1v%ZY1'ん2-B'Ss=OJLCDZ حKv/ FRc4Zi .wMGJжhzFp8&i(FWi+$*H+4F8D "3m8+QVʉwEi箫HP즨+$HWأ]1.y.)Cu FHo@ ɸRtŴO)3;j? q]1-f 2%g+ce* Ci0P{Xw(13]]颫V=WI%FWHӂ]WL}uC-rtq5MW.0e5ȃgj@9B-I2ל9ВbwAiв{3|oW~PCd`w@&l]APn']3L3ĴgrLIeicvMrrv~ㅋ嚛e8v!G/_eeU-ҺY`jCK]W>jVw Y4)^#Ml'uו_0?1P)vp #uӮ`S6@5.xhKިִ횅Bh-iZZFo~rnA7VU/m:? >DvIW_g۝qk񽼹cNqjq` w;A:u-7UaV]*li=F]62zٮZNՂVV-[ ] i?z~tFu/{lF.T[mIw-y*g\\sEM˖?󶝁Z>3n~m;}_|o?LnOMJӃe4oN_Ƨw18w:5>}<;Ko6S:&ZV? 6\3t*Oon(C\Lo.x_`Mu5/ctrEďG+>wtrz{=4FG-3X3盻OggzǶ@:W7͝|;Ay،SB.HٟQy)>z% -ΚaX9k7- vHw$t>N|Ш*ⓟ;x+=>{FqQqu4U/e m/OΫP':Tտbdcʝw{1@J'wz7(> ꛟoH\?Lq^lBSfJTڪDdcЪ9 ;cIB:S*Cu$? :T n4UyF.Fy<<Pw{& *|'@s&AkRIMv-( Quㄼې&51˲MF>߄b&tжޘ("uDw>R6D1WBRVI| m)ݍBԉИ0 IzI\qjLkTSv,X3}% q/*dZ4X)5&QF!k_cuЏ%:Ar,#naD:A!g֥N'ɒ ]E`qW[)bڱJ$(<9VE`Ak)< *{]1enۚ]EW!G`rtŸKӢ]WLYg+c՛e0A UStFi2$]=B#HW ]1nPRtiAuŔ-+ڏ|pSnV!<7kb>'4JUXv|{ĎoT BWm Wii mwZ6o:~l-9mm-7D`;<4\R5٧9L%nKܠ:26Xp4Vh:a &w GJhxF5As㎽9p-uŔV]MPW<=I8UE]1>b6EWѕun3yꊁUűOTO4J*uܞVw+FW+'dZ^WLI% ]1UbtŸNN.w]EJTj *c+1?A)3Xx|Ǯ U9JtF2{3 ¢V=銁þNM>?mn&hǢhs\iH:^}ǃ >>0ʉTבNiǚ;(IAY2d}.AO ZjIPFt)9%t[谷x{嫛*h$.K{k8VVfiхVUƘe`XV!,Uݰe]Q*8ª^/ Ik绐Q+!Qw|1x/dL(w@ZP&VJiڼ3\'=RI䌓33Nδ&+$Ut5A]ap9ADWKbtŴrS:*(| U+{7$ZGWLilutt!1bܽn:HWLkU(Cu5UF#'d\9bZsS2v5E] tTPf2^.+5ٿdJg+ڱ@?`hsp 46tC%@fPU%銁+ N"RcuSӟ3,^ŝ37H 2qbOE/dڱS(s[0_ϽZۛ򒮒н$D%@ t gTq4iM2L6LE{4j+vrJ F"mGL%H$5HӢ]WL'j:Y銁w^$fbJ(u OFNtŸDRtŴ>W2]MQWޠWf10]1]1uŔ]MPW8$*% j%EWLK!w]1e(Cԕ`k`k7cZ>dJ[k?](0}맋wrQw%O2kV ճզH/T??\Vu׽͇BTW7]ey>^IWV:4M_^߾?Ƨh9n[(w?Í/'g'|s,^πb=O>(XPW7}ɯS\Na>%J4j#Y_Ϯ>/.N:[Wԧ9UlX/s :XZ.V>6uH|x /؀rY:˦=m+Y~߬#'~W]3m(Y{y1ebjŧcJ2xxd.I>à˒|?-ˢ.[$Xl&*"N|4ir,[]µ2O9/`nj3!91hܷE~[ȭt/m)0FY8l݃o/ ed[j,F*@|T\FNy${ש B Źc7L%k1ݟ.F Z?WU b5BU5fQګ+{WCA1z.AboJO6ܾ9qN|k6SѠz6zoߍ[ItE{N>!n.Ere`~O//Ϸ0%4??>hS>?*~_ss67(f*uAa2ɋ‚n1!(K(e%#.TR ޸Cni *5[2Ā5)8p((g҅Յ1`CQuMp=#`] WD;mRAnpT"\om rθI`npHo%/i%y#Kⲟ[9qLtCcӰeAtC٬8'$@R<nFŷkہId/'crLK.SXܙv4VO6f :kY7ْmeX &֙y͵pĘLcQ Eγ(SnrO+lV TܟjFAzMtQ1kq6V2龬V4]zֻ&=ajK[jt?DKs?u3DLY + #]*DI5JRe߳/j{ܽ.b$*&е&`RV"/ )T0FJ$RHKnUXi䲴t~j 5K,iS|תXLgg뿞D4l?a8Z'?'?5KH]k4I)V+I.DsSw}/ojoeQ^Ce^1IxCc2>ldOlY։G7Cnn"د  ~mQڂ=,@*uON Z%+`T,+yi`ZPejO !sD}o9E&5CqC-RTHrqYbqȶC6c3U>c&Ӌ☕ ab0ӼyiJj50EQ9B0b㘾iNĎIгhV~,Acv9s\T^dZ bۣ [qVS(B&mO^kr@wyLc^9JhZ o%1b2sޮ/)vx8y.au;FJ'U7Vfq]JEq7ݓx/KqfC ;gƖ?`LH0v:aq9| :;ůuW63OC(PxE,?p1 Nđnõd2NEL$)p멹LC簮^  - HwZFҸ:|/9~Mf{B&Ʈ@S<ә|x PSKND1D2g e8(A8\ Я;ᾲ.h ߗv/OG$x=9/kD%2*ULK4A Rg5RI:kdbbon\_Hi8OAЉ.:\}$J(d/@Ö#H4nh\i[JLB&+fpAɄ`Gm1kzf "Xę;Gs`DWhi>oeg6G7P~a- `J 9]~5 SӴ|&_ZH_(( 3PIhȣx~V]@)-sB7G#B+2mC|%]ˀD3FWa5@sHW͑Y8㢒ۖ8~{IsPx̄$8:2 CsNݴm)hq XC!0k2EL C"e"_JImyGF%NW}RoFy-*]mM};'][wMQp;Q։MqMy$xW{oqwgAnAit $#Ik @taOO҆!iԞ44n%P\rAc 5Jڸ@W(f$vM,~W~B;=RRz7AXjMr`PEUm*Ccf/ή/`Dk.| $! nC Rx^x}o! ˆR+哵Ad!/'WQOK(%H4xxa2JeԥJhi ɧƶٶBzTX079):ר߿&dIQ9`XW u-X\=yQfwù{$Hi&0!XM? Y[kit d)BOpKAu{ MBlv aݼ  HP}{(mkD@⨬sVrv_L4ٯ!#q9A5H ~zߖo~ J̑D)Z/kr¶kf[ŀ[x RDV4_R?QAǓI=vIoqogAxp~qsV̎ϑ QO!78"Zj-0Eohsid0ÀI;ݸ7Wu#vNܴ $)0 A%&.&$[ຯQUbNB@&+)觩:,5CQ#_%*|Ȯ{U;q8>#CYkRd}[<7]-16Vhv!炠` 5+\T 5pq+r,8PXqx}SǓQy3xc,닧mKxG.aWǼۂX8xn,ʨRH41jЎw,xjl&ޮ6:q{P# }+<\eRT2g 21DmK)?v(._`+cX-Iόk -hxLJ00eaRQ@?x))9N#f{Jc1ӈz@Nj=62;Saè55HbzJ\lyGMt}H.oy Rl,6uU(+/hƇQEF!yJ+e!>@H+>Wڤ Y,')+q ~/Jg*Ge>t%x~1 ٛyGijU)XJ1ps7@Y>l Ksd! mꩂ*Nʾ9UMp,`uo В&i*I2m#5q.,IY ˚h*<uazaI8>BEi}~ͣ)/x9xԺBDyTrʹ=EOmf]ǤQR"#HbBZm[?9>YTvY͎qk[4q\H%$F$S5%9&)`yZr֥ =/n$N̦6iRm`ȋD ckM*xs-tG҂}{X=9r{i6Vh{Ijݤ<)_0)OyVh~/R h,A@qid6v"w\p]Vxްy*EuVkTߑDg?0:|~ (ڹӡ=Par,~f<}@Y@k6*w |p ZU{)t8p'%AIN3)o^!HMUBAw9i\;v$YCDe*y9؁A/m< l:Q; ERx$q9K.cV^: I.;ũTƋd]8|a5*2 hj`)d'"Җ(;Ǫ:}pR+ U(c'ֺN.#Nnԕ‰G[uDoaQn9el;FךmCy.UG~ l%Po'0Ȓ2ֺR'[|pF޽S]ā+-0'ԓ0ȹDe<{ vyH?߇槓R-0 |KY>z}m(مT8eN[IRPR|(ܥuR[*J\nD}mLj+.P\c+]F)rEnNEnI!z)ƍd&bRrisC@mU[G^ܹ U-QXrXNk\wI„-9sVIm1lYu T9".ө#lG(ʷ&9&PT@ H͑U=I(V1 jurVuC6owOVYt9݊q{^lx5o/d-diɬ':YMCY8u2Fcn?D0P^Bl>Oȯ˽q?WWxf-^\Govh]26K_wm\hMF﮳<6(K^qsv %@i"T@:O&<_V^5Ӧ$V?Gw_ К5=UVl4$ԅA&.X ]EJwy³D?1ԡ/{I32i9>&pY^Q=cc*Qn]Ŷ gkH,S Jc0 mK;UDPGSG9:}D(HVw|cWM@1&F$n+\ !(Ev.&0A^HenhɪHuqHrܖh`0Ad̤~ J`r5`6N&7R)O~'SU}Cy4\6JbZrǁW)JpH8VN[$iSA =zm%U=ʆF\l**!ϳaϛ\)̈́֌`ŧ2͛҆"PkbJTk7㎾Zm{n.YDsߗe;1 Ȧ)b/mUի Y,aKD9ZrH SA/Uvi H|]1mO1mxUR3^u#E`yW>#\ij; 5oc& GSz[2'UyRN,c?@6Mu[v:Z 1/'yXJY5f>f_T{vGY:c<)m}T"i/O[UWͦ51ï~AY ]ZU -@ܦ\Nܭ*lqZfE &M'o\֭׺oN՗{O4hRxAykNN#y33ŸeECfjsH9ujFbiM7_@buz)ʵB({7 HbL39ԬX6}I PEef!D{ ~7hw0,󬗝Kaٽ^:crlFkSn܍N3~"x:dF4wɍuoƓҰ+)dU4^TsKd<i/f; ۓyzQ]MYגR|(tR35k\ʑЃt9~q.N*y!W`Nq++̾:}*g9\9|OrPE(M Җ< }5&` DžEwmRVIRn!Z=욫{HK8`u4oYq^( =b՚ ҏ# !iB )q{zpO/s[u UɋGU J1^؀c΢ȡ5[gw$$&7Nޝ(^bΧM+?ZTT(@*Q4o eR.ZzKz$ s!=I )0mPoXa5k2ވ浉t> a7/~62 hQ]>? Z }?xTAR>RA H]71u<^̳`[|ƞjN6tzvU‘|qg~WW#7lӝ7dlRiV"˃-eCgG  a&0$1ϲ-<X8(Vpp0,w`VNdv>(Q Qو`)d̉ C".!Q8"Mݛ0~>ʈ_x0cq'z:ֺ`Uk H,%gRxD3 %ՑL԰|8180a>xpŠK`)R4*QHܽOt3݁2'rمH;C_ys-f3XA+؈݃IF0dnQ)Y+R"PkJ6LxHEiY1g2bAH% {t]2|ZcGA21HFF&.Zjjj4c_w{ns{ǣ"?8?U C(3 OÕ!Lt7NY^*]]Xs$n[MT<0j8FmBC4thm"U"><8cD(a8~@da/0Ў3&3 pZO(,3SIT$"սdzѲ@A_9ohf([kf+|,m]I;*׾9lz V Reg$Z Kΰq9V\?oבmV/@L126IJQE"Q$61;i~F*> =*TI:iQBtNIL_\ŎC9) ]n6`6\O8B؄k>?<(%FJ|:HN 6P|@]|>?݇۞'P5 iF95k|ef,KW7s&x*E4xz e%5e݃u;Sq>^^_n`"E1RR ,%LD5$(.h,tD4BC;)0%@Dj%G Ay Ӷ"y5 .V`J-7\mCBW<3fxA;z$/Xh\P0d[g9fc\_U"j CRe̓UwZ+nL捹ZK 6Mdӏ%%hL_Yp7dj`Wތ@p"QS X'0ZQ\57ؑ=aZ($xZxݟ\0 ɶvj.y8j se|x Uj׼lC_m2^2A<8!G7-@RcVB@ꇣ$ DM<<Ή 'g(B*F|޵ Ic"wv=W{]Vg zr"!~걇D6Z~)6LFF-e>8uaR) đ݉GQ6L8˪YJf;*&FN +ch~75xOJ0ߌcK\fblL_+mE$vKp33L5X m%Y-Kb{NiiIHɴӀ!:@qe]F2f4,f7dF%#f( eȥRjxȸLsʸq«:,__z~齼?ӋxWgҫKK,M{k ywxe6 B<>g׽M[`K,4/0N(y?}1ry{;߿_?BO1],OwXo>\`>`e &\`e &\ ţΥ K*fcsXu@pnypEJN2|uqԍ)4Q.ɒ xgs+ETxCѢ"vVѥ7VB{Ę*J5s [AnZK?IjKhδ Rku*L"a)Z (j%nU:IӌJhFEY!^LbPyUs\Rb8ZGWј\kZ$ͭ IpZ)hf@gLSz 9gƊۇ6c vV ]vE9pChѷ %4>œ M>< )NZX8e9t=<ƺ{#7^q-8Ō*;/[ :XHZe < bYD6^`U1F^4;Y,0aߥa䚙Bg[/|XwYU\  bdƱLi/MkK6l)VO: l`AF <7*䖂:492SPL]5N1$.Ьдpkۡ}o7A=K?b޾W7m)h^5wۅ ,`h.2 '6 x`!-g>9*,ZbV\VTHs-XrS|aNj)iſ&du?Nf,$p1'$)e +Jg#<{q<NGbBax'7ۤd8'8m\ ]פj._XII/.| 7 EXyl&RKqS][V6M~iJ= 6"ez *Y>FY_@ RŅ;SvEήY-RjjSJ %wgú!^V32͑ȅ]<^rl'J3BϝG\ϹCv ڊ[Njp GA͞XhEFぽ%4>cAbz=^vy%7e]"Ӧ%]5>Th]OJmXOR:;rtϲt^@rHnqnq9-ʜsFPRvW ř9fq5$)VBsBsII^Iu4r~\=qkSKh|H*9@MR@MH%] SSy[zpaT]"ƷWq8Wx:\2qڬ3ɫu~,Wk[4v~PFzui?=OlmNŔaZ&bXAK/֢ɓ僷LT`bQz2#YEK.h\jPIpWD0Bgeh\pCp,\QyE@LZ-)/*/mNR.OXPukiמ8UK801u"+=4< =1P{D&ѹfI[%\P,SNfqگ;M͉9޻6-#{unйIp:Eb!I&6 q$2PzR#_Y`A+&?н׺hQ%ٻXbdaljIhm4VEM~}?)'K,w ̨NuVsf; w[˹y!dQ+8mkIrմܬ;44ӄKXBW$[ Ei٤h9@ҚX2 9b5hmhH|87L,7 p1^ Tx;3pN{CˆyPC=d0d6X8hrzd0'H\sşoմVxpw6*~mEzhu:VGhrālt:4BVZ [B}[pbG6xou:Ne[&ƲF ֒9rV*PBf pM8MXsJ -t.RhsXz+;OEI}-;(Mw8; 'oZBpB ޞ 3FpVfr)3yȋZvºew ׃Aʨ$*PHL8?țoJ%CvyF bOYu5~w_3Pyp1a8Q %VS>.cSG,ZcKuU^;эU,v۱v=93Ό+L;IAFQt;%]Il07#ҵKw uA2TyrW &dKyLW~-6侊IY52> K,AfT00 `XιI}:vEDUer$@IS,Iix&o#C׷7I\ )3jAKH.8Z[;7MQt93.TνTpGc 9*df3X_ 4lwV)ؾ~78$tBiv(b 3gAMʂ FA6 Lfo6tui86# f{P*ŗ򷽿@Y `R)S x#;)b%4>2e)X*Hti%`s,Qz49r3,Nn vׅŽkF }:d t.RhQ(ԴB XF}oB7_|[ [l:u,B`9,՛zn.H%8?fy_ҍ*х˥pS^q䇩pJ[VQUFVRļ)}DVghK2Z8d$L"Y$F%4$H'@%2GO@hOZE9Zxȉ{>m˞hC7Q .m\û64..Al%4:5 >HӁB~t!V~)!`ԚDe%4zl #yORN-ۤetnSZ u"VZ\҉]aEOP(I..t@V/gyCKWV"XHVi ̪ڤ C5G'5+)Ž:ݴ \hbcu >eim[X[V36y}㳱, Bav~| +RԿ]m?7QIW1܇#VG[&9+qgVT+5GT twlM^eKBMՍW!դqwA8+LjYK٠|B0)~Х!T$ATb"pp42n mb!U|%ysP/%ߍKd&Mhl& 1Nrۗ+q0ESFмn0X1=;l-Rlwy8a)e?E-@<(JD:;m8'cgomXU?+)MqC I - 7M]\(<&Y,$6AIN%CHBh OWS߄5!)*^-ۑ.F _܁`#p;$-|*}s[?N*٣g>u&<_b L.>,F#ؽ?,񕇙_M@sQaF8)}͠t M*g/IM*]Auc%mt:΋)k; W6œu b7/ M ǕMi#h_Ϻ6`zT bQ}5dMij,o ݱ?t"ZVl &[mB77;a%c>o0Vv{ߞ UV.9Sg``J.τ涨6/}Mb[)+Jp蝌4稏 AH^i%R M%CS-1f8 Pz.ʇk7bTq_pJZKx `Me lt*MlRK)XO@D!~{hύKc,?ӗ7{ S,Þϗp>/8XG.۳0vj?WƵ2\#=bL#9BszMGӜ|pPa;'E6`/l4)uPDI(y~+l:fqf\0VVa0"+>}ҜX\6^l[ xwp IZ` )-F]֑DZv ({J2ޓiDIȏŊt<Ŋ7zh#wYqCi+hdǃ9 #:3zoy*XPt~Sy 0ƨQ r$s&x9+!F =Wk~GqsoE4 au?)$/-'M坞,pXqų7w^$>eO;2Jߠ!;l")OX7mdTI jWcJp΢0btySuLЙ8܎d}5β__>c}vc'Vw*!b>?՗~%$fJZ޼_'oB(kB.08*4VI>u6P/jWr`Ycޘ̇~wGB/>V+t[FxĴI0Я~%eZZoZCgn;#9qb%_ +DsW+\cU@ #>y7zĴӷP=;\rf43BnairUqͲo՜T #9mgzړIeIػLRW(~ΦIC&:.$3-dqE,eʷ6\t$U0kdC)P6 *4HS l ꤂m@%NE.h ?[-~65I:3WSY\s$ln2 O^;;o\/D52vd\^I*q&JutCp]Hs8Wb~d i*'ԁ [sbx~[Oǧ-2<+o|[\bM'e P"mBReye}(ޢe;"nwR!\4$-{52v| \ L+.Bɥ b\ 7. FTxU*%L.*]q)W{#3[ɂd{lOHdOu&t{2NJrfD"7ҳ\a6q~ P%dHVk5'*V5>Ef;t.|:.dc-]b$!薖4uu2KZ\F*VM̒6:!;Hr:Sh kd lXm }u ]N:No%,odR-/IWdМ (iR!ktT{ndoׄo=,/ mw_5ެ5wF̏tܬض cÂtd\I!%RCxCRF52V-%T: WV̪wRnXAx)ڙQUy 'Y4IWmkKҧiŋCWI>*Ӫu%$u2xwÝ=(ퟖfx!:TRF NVB0]7#棰 NGEI[qF9VhA~BO?aFҌ]8B`f38.sO,}Ju7.[>x0iU*7SZ3!5rkpXVWeueFV(O@SسM -Pنc۹j̝7}Ԕŝ_4'z[5 QkȮ uw24P tJ_#e\fTOng:, %62Dd=myDX(}r?LMtMx@M:Q73'CWA#1gD6 G,o‘8CC>4Wcw-u1wIRo:$kd|4g-&rcYy1][ٟtcd52j[LrX~|jmjƵzD_`w] 'C%cPf*пgbR.م)qäTY8DRKT S2y%-;*a߷_N!Rx%< UF΍+21JQVe I/B< Iʝ4DzWwll=֠}GNķˎ/[ zMlERg1k0IjU&`׍yB2p-"@+HcDs!z=p^Mat_Tўl3v2GaXBsSk.@Zk!v_W|RJZˬHkcEٹ"1._^616p/`!:8)X>@bH4wy[ |E2L^@'ŵJlnoXw|h̀gz̹[e .4BФo.J ]٬ADF! UBݞXtV  | !|~/OL.FuzW,wQ{83Y']NP sDGLx5gG,\(*D!`CXlc1U, %d3IE҄EP%CT jQi+|s?THݩמj&}+\X~D~|\frQV;=;opĮxl"0}+ |%bo6 مs`\VP \`BzOrZYW52EX\+!ZQIӜ}6VҁbSD+Kf*Rc4PO$$VN^:!@;e36JXjdUh bJ*g]=$J(1?|bQ,%Yj cI(wXXά |gŘo{. fm>lj:j_SP<UԊB+\(8J1C~}vk`F>ծu{ukBdjCO?XO[HŸo:t%J61bk1*-:ga\uu&RV~ YSYbG ~#}>>I>S52#tNSf ߠ[Rk;dEh-I+ Fe}iTF0&:C21u C40m (kQ@sG##R@c0(GuQo3 Oi Vx $c.FwD%ftM Cp?SUd2k% nD)Tٲek&DT In:fVVK"-l1P \b0ApZ=S?Lw}F+gN b~`]~ O!@ LHe;Ww\IngB2|tpG24rpmx91 K+# +˻6W)r"Tեmp9MFƎ2xSV5> %.+1Nr Q RyG%tæ,\Գe*hFb@{FqrʣQU@-MMuNYDYy ;jDH/$6`T^x ~R@sA-y φh1A\EhX"*MΞαw/Hbt֬,J IL**Ʋy Z)k I%ƕ$i$_G%}\fRc<\~@+0#|VXIv!gKŀpVO|#C~ n}nh52RwzbF- SR?{WWcJ}PRm}[4 t23_G)?t8 t'uEQ9EFe>U*5j䖄+*{v{ɤc9|6?f3یwΐ*yKe&\!,_12]f}-Hu)Kq^lV%*i&;]XۨV qm[tJm)6<™5IV<lh8v GDO| g81)j>*-od7i!5;gSoyT]DګyHstP-gmQ`wƘj. Ea%r,(>'-|gmdsn]U5QD"$jϖ혓̯E" Nm7Ɋj2Jn7?aDzVab2D0-@ʱ;@2\7@wmSJd{f_~4Lm@9=TR6%|Ada C0\kkžfY|ȵd:7 i֞CHXj47VonMQeza?ns#ns V?}iS`qiǹrK%e3""#c(j[qٙ}d `)$m5SK Wd+c֏xަfZ8Gȅrm'C 7cLqE_Dl*2?JlF2-)߁ĵqps'`|.hll[A{flizL*uŨv{Hd^EI,Ks`(|Xx@ZԡdG/Jf.HfcGzuy^sW(/l)y iY'ׄ|\Wp[ %R1LQ -~D/\-}:6b[ -O12(?M op#Q-'Z~k5cȬ.iQֿO=nUhh6o{_CMhz-[~B퓃»(%(Ų.DڑV|]tW[ )5 ]LqZJl|gm?;j^!NM^q=CGl'r@ͶU$B*]Af s@ ^ߕ-c~)c1c-6O-~6}ݚq{$#ga.Bo]n(K.]ֹ;gӵmɘ f PG%IsvNJO`3v}q 1S,b&9;xb9|osNvYv%MF帯5Ֆ!?G`d˖[d%ꨃ}މGbhJ;f?^;߽MS;H"$hGnuğy_;8ZS*S (YG183VmA\/- w.B4Acڷ`9n gar(؂#[p6Oȏ2r?Spq8OĘ-8rc#G738Nldoyw3C8Px24,㺫GPs;)$tܧa$3̽'LMLJj)5k- 6ܣs<3; xEwߺ?JяhO4Me\H!"v?|# 7Ѩ&`v3qH۸dB)Ӫ] >1kfd_ *{NNޱ<|pᱽi&f"΄-I)[|'S0f-0t[ Sܷgz})O RٙTێ\Nd5PONv[75֜{^ Mz*dkM|O:V䶜4%?9юȨްQǤppFު:A/^"7`0&nr㖷ڷ`i ,x?d'nxNG56XTbL5D;_+>췂QH{dtjxB mܘ).}P^ gш{:Rڶ]idRbSރװKL'ou"G# J:qɆ$.hy$k@S1;Ѡi6O:mH@0c"Ex7R۸+bTP0mIbc4nI-Y<}aɏ?VLۻDX |̨cqϰH(jMĸJkfQkO뽼 "U|a5ܚ~y:cHW6o6R,dQW\+y {Wh~ݳvKrJk)d)C SB%ʖ ~:-P܇po:5JE9G6ђYb4kzeŊ)Ls A=T42`ܨᓵa+`||'p$ޯ VZno93H 4l"I墭򾷞bL1Rl a{ 1oiazw5:YL) {d0jMIokgj=g* g\h%M5k*x./ȫqA^ j\W䗈J#Œ.#ץxU)RJjTTRگH'k+NIQ욘Yzq6R089 s1kܚ(3Y۸M0jRF.O5t ܛxֶ1Ly p!u>KX[ Ipn6RXX_~IR [h˒g"FJD`@V[@E vE?%aȭ'#^UJ)ujԩRSt6 mJDk:^U5:ERj"6#eJaVj!3:{szhRmqm `X4@b"7-X;:`f8,dW;Kᵡ ͻ$tgJ ػ|~Fj_%MXp/9-9`vըF]5* O֤ܙWq$;p^z!3z?a6DZ@tW)Z ט+'hs "֕fIWN XC(шfd8PoL ?~??S`u ";c$:58rwvq`dDc1 '1EHS1W~\Ib?gp?~_ߞ=?K_X^^E˿=#<;/~j<گb> *P}*zJ+*F&(\Lˉť ~Jpf}utX3x/W/mGTs;&Э@'a*.gkZF^b(N#gE j/:$ bO0wagX:(y~9B&%?dd@M#t] K/.yV %@]A3+ӛ/Xuɓ~,@? ,+򃣋`J="2%PhAyAF~EȁP d9Fb`~SΡwgCtk+ Us߼_W#<پ?'r/߶tr嗾g jҮ"kv~r/H27z\놤WO2z`8;V4<`ZK sg&%j8+9\DXpFr֥9C.1hLFc"ߘhQ|ϔ,d=Y>AlQ'>EҊ" fҤ,;u(BO!70/%c݅.Ep~zF 9'=:}{-n{'}Wa'd[MXQMˉƎg?b8fQMZRն&'-D$|驊N@K.Fޱ! u#cId\hec`vV83Eugs \i F{Ҫw~>7~0F V~$pVe`.k]T,ޝ~p@`՝ݸF㛅It7#fҎ',Dc3+}w*B MZ‹6̼#I eZt@A`ly< hfp OqTddJjE]nn.dtysm0n˲m⯟CYKRe4d[2эck?y K $аP)'2quzKT$9si&խ]|3:@ȵjDǏH瓄\oD­ץBLi[ 0l5zsДdC,HSX^Vc0)9:?'}kҗ~kök35l&o GD.8@rv5\3SzX*Osyٍ学 0R} x$5$m>vp@' ZWz?>y)>,{*o%ʫJij\i |ЄYSf1ϲGe-;pyel\=A΂ f63~'.=ld~~\5Z;ֿ1qL7~H-y r:hO9_sNG}@s 2̯)M-#ze&Mպ犿-]A7ZWxm7G9Wt[FJ$ؼ 5ޱ!hx\xi[]oLռ쩬p0̃Aɻ~g-j5Ӈylp '=&hW%fx֐l{5\squw DGz]tڗ~\K9o'|;xڴPNk$IN?D1behJ#hc#}W*%>\[<"q+I{q{|'V!+ӓ%+B$z˝À 9h~zx#MًFQgOٗNHb&Law)IעԃP>Acl{+ ߭O-]/ncK<Z5e:O vwX‰)U{C'ն M"9zG.2%rt.>Z@mpk\ @j=;~*X+r?;r);gO ĈO,°Z:zGw#)k!%M,z*TƘgB|,\= zxε;~gUٝ[\2wQ,!!bn%B227_xݘ,K f߸yl!^ͷqB,2cgUv>ѧtx w^Sw쯤[^7w`~˝tv!^.Ʋ}y5kfK~96z Y=lMǍ;&N.F)~\_\~( 9]%(pyͿ9OwJI_?.[տia' y6|V7#? :~!>ۋXkwNz~ͳՅuoKԢϼ뇄_ҐIS+vUywD`9Vaxqߐ^y^`1~Kz[˴y |W7~EA*ɮ yp$ '#ny$ρ[HJB]H%fNY<ǖXI [*mK }q ]O2ힼU&ޞaG I%!d,#3I?Wۡi{C@arsȭ! Ie| CӾt=nLz K;1O7 ~ܩl\ E|HjՈ1gr"~gD&-C~&h=!#Y ,cw\5%$GjUhok$E6_W--X{5C0`,Y9y!4r%_y\g%^sʡVy91jU!ƌlyAS@{6&M)æ]MΌm#Y6bSH{p\@f%HXBk,#$Jt[]cS(h>bE՝ #Wf]xͤ`'dF{=r͝Jk&NVt Sr%Dµ}ol6[FPl5^ky[b%de=';9+S9`7oyӲg[.d2IJ!VւNgUiƮՂ'繨U5jՊ1J+Σ'jW:Fb۪fǔ!fsR:ao@( <xmDo&31krl-3zs>MXΟzDjD~_h%h4LPM2`FbruR|H`i5A4iZiA4UE'!#(| ΑGcB3; u,O;y=G0|xʶ AynyTƬ]bp\|0b+I,>3Q}WZ˵u,'1J*bAa1ZMbu`ߞ;`ov (;?Oj+aS)%r`!~|>>ĀoZ𥹞TgYJ`@@Q5L~z?.Ќ'7 f(vry(^qsrqF ~ѥsum-%İ>! 48=UVr g0o>}y)O<3bJ/ HdQK6*D!Fs}Sz4s}%ĀNK2xa!ko1`/:#Vbk{wpִi[z41&Ȕ9goָAhTHCl fPm hff ]4)ښl$ȖUyYW Il=x3 kqfug)ڃo1_=AZ$aH2j gq^Jw*!n2(6ttZřƏid?'0If`u&ĒZ-bO"Өp/, %NC"|ԺF({sP ʈ6xSw|=[e-Tu?85mոD5I(4ӎT;;21f30%XUa$ y`gBy$Y5.I&v3P ~ 6!zeq/;bN;=ȽYcGG!C91X0Uŵ؄Ҙ{Lj~'Wci9\EdNJ-zDsq&k p'? sX؝*mv+6\'J| =̯)}pu4+äLf57A 'VWԅ U:9(;; zu)ߤx~~ (~Ϩ&dv5ѦįuA-q^z(Tv/vm;'uվ3R!wYx`NHpRܪe8͹YXpF$,n1W(;ەʶVm9}݈6d&ΝԐLֲQPq`f9oJѻj|N5; iB>hrXg44N9`]rsC~>׷;_6ZN*֫8f+0%)4_KIkX*',ċL*}?V?f!$2;MҡGZeQ򘫃u؂hWq}Y''rq0ͮ$NQMUf.1OhU1.ƾle A0y^gc*Β0㎄E-ۭv5ϟO5|wo?^#ת4=|]ӭ.!8t]C*ئ{f N~Mc:zU{~f ^r?1;G.ꪼ,V&Й2OS:OkƁ~Yw߹g\j3<~(еca0U7G}̣At|8j[@%XH؋qd =bSb.{)k Z]$[cNy?<1zqn) }cp?[rløh8^zP4nR, +`vƗI\3O2Ecvzu7eDR{Go |bj5AڜumJ 0y. @7O;6<<>:6sz%FH@>3,;2?+d,5])E1']Nƕq8)#-Vls.?5 (.bY%Sgv13eݥQ7uMG+CѬP ]a KAow4*I[x rRznQCcP)tkȅg,OVW:uuѯrܭG`.-Hӹ^5#Ʌ~$p*VB﹯wD~+XޢK9j$cwo )ۅ:#-'YoYvg I}8R\J'faǐ Hy@?y\f FWz}+1\v ^] Z-՟46[ybND\djU-)v/>79=Β 8 !?^ y9`\raʉ1(g֐KT`8dUbC.Étm!&bi+lbZ >΁-kug/\ˮ%(j>#DY8KI!/=熇_Ԍqj͗Cvfy``X1pjB6~&҅;SOya6t]fW(憭#i-pDlKhu{lb,(g3@YY[JE !;좜~M 0.E)yn1jse.y GJ%H"W6 vn w DUHEӇ>^O>:4%0U hga>6lߺdc*^RYg9'LF#;|rSM$~uqNsN]fJaFca샡l`%XUX VY/3s7/DI̥TV#nW< *s, 9)䝐3'6l14Ud 2Y 3X0ɋH'<[d+[VTM-yN3dle*b{Rsk>i}j˘U`b,D43Um: -@ޢċ[ec{^D05gѲOs2GC6Ru+F_YŰؙ#{#hUd4ƏJ[K0,DA7۶z{ӠK,Z2ͱ|W?t?\m5^d%Rg{0[ E =Ӷ.~ieF*[$2?J:%F/8[:wos8cSmN]Yczي~ckNɿ8Bhju?n-K&\Bc8U!9t[U֣"E:Stc?mzl.zBf׃4?ea:y|8Hgz\0<,nлҹɷF<|$;ָs@CIa2pJgEw?!>s!39`V!$px&TAqqC\·h󠝇vlmDH,4ۄ .rF!}T1",𺾤mTqkȥg*{3kkS򳽷U{_? }YxC'b7|6s);Z٤ZǞ[}">X9VnV}sT\13׬]Q8O«:y'59 D8&lp\!r1~=N vȋ!sjO]u)M2vAv=;n yh@5@A.{ޘm&o"0{8n yLB';V!( ]QNߧ;]s96#܁j!/B`[J~^13C]ajv0x*[:]VM9soѦc6 -SoF@!YHR&$qM 1#7PXPʻ5$ҞXzlxkړ`{㠎"87ؓx}{CWC8VOtt;1VO6V ]c6r\āhco]bPgNl5Q`oܛ]-bUQC jw@vOtaXó7Gtn؍ϝ4kvy"C fw=f[fwz,ts f7"yhܙM#ڏ5cT"]< jnPGGEvFl:js1jYs9a0ݠv|`PgNz쯀k<PK jwРvOtAva'rr ![73;Fay=nyB1O; }!Hf2Я `vaΛUQ{0N joøܩMc|MԮެyBhjwgtNӶ#_L̾w| [oS5N)ѥ$tH9E0]g.!ch,{*׊5i(sIZ+)"Y&o݌9fYWY?3 0R`rc<bX_V-Jsɹ,!ӳi+>AۊU (Vv~GOcQ3;)KX~o=w-lMۻZ>;~;^zOgZӢ;o[x|֊ybJ~8}wۯWo[3FL E3w rr&ak[Ͼg| J$0Rvg٪Koܝ|sddN\ d[Ԟ?q?MΊو.ǐ>gXkixoӜ\.\JAFrbޥlI%r|w{aݓQnWun7a7UI=zHQcK~c;Sl`R-G&OgYH%,`!i팬g}Fmw܈;1 *[X79)JgiaB6@33ΘZ* bi9g )G1d.7g+Kƍ"Uy;dWO{MB4F׽ {m2[ymm'̻7dKa9O٪q9͐~ԕ*N▊ofq !<hI.v{\nʧ3ߧ7%c6K ܺ{@_gZj7{A!]M,dW#oeRdFhcd;1NSC<5T )5ˌi YxQM З1jl?C&βUF _y3+[U']$h]wwnʋm|^w%? >b*u=&v11% |d;HqQMK ge~_3.1%&(]eeVt{<b/T 5^w #1(T|0}暢\KwWm'T(b_I#xc+WNN?Գ힅N^:;`žBa`;ٰYP|K J6Э2/oq>@)yɾzlIO!QOq 24n-AӉBl7le{DV"퓽wGvnr>xh~0q?_(,.,|*MzŇ.;y|M*nUT$ɭ]t֣C]~Zf;` Ӆ[|6hXsq౪3.hBhBSn۟ h=߬si)2p!@G2;OJ+;}1^ߪLZ[e7ІBV,!17=~)@zvxhá pF-+Qx5ysQJ0hK ,7l+Oj'jdÇD##aG&jEdv+ Jvډ:[` ڔe0G rJ#vND*]0O!ߏyCTjn9y v;B#c/ jN]?EQݞS;!Q~z7 !NrCCdٮP;4,\]0uNLB'Suk5Ԯ؂QjN4]<^ )[}b#mSؠP3FYuQ C'dMbdUj/YJV!C"Vf&ۋyL$=bVuރRvjAns:P朋yN]?1}vB; G !ayCYJ}LUNoQYOfy3=?JԢP<}+ofΧG~y]p%?׫Ajy(ɮ;C8bh!v}4ϹŪl y(E\;fy#l1,Z<~3*$708xi3OR}u1 NMm-k,7B͹죟݄UFV.]edGގA y>"Fv'Ȟwz/ga=Ra}vf,,~i=0f6#ھ].ܽ{ޙ6#TZOuTW>S]/3F d"iS;ŁЇ'ЊKjlq0asu\`N|'yEfc.!m4HvO'FAt>iYL F@:$(Ȍ}nO)> }=Rh99gl4]qDhGa.:)ɥr0X;Rv2dvF497O=qo[Pbn9ySm ؽuDҋPE*Ľ"q0qoev}H3;tRULgg}yhRx@2'f4 uьG(gF.Yo(N]1P'7=ye44hZ @C6,.@)yp*]}V!ï@M%Y_~n0ޜ/4YnlYKzvh%!пYY:wH'._o4o,jDňKNؠ5tr꯳%fE&v{ GZ"JCx]ѩCY rc{H>hefM< pb,A 2Hv\ ] ]gzd=ڪ{Ҁe:TE*6w`}@Ї 7 Uh'h1ȴ{0/PQ#CGٽCvN$r=$hN/l!rc2.sU53yS-FdB!ZSa/O0>0a.RxB`4LaU&L^wFVyLw5!x_,Bc".q`tX4׏o1 G|NA9KR4l8Wgi ~8,6_ FNq_U׹$򙔸F&ZNz{Gy~'Bh2ĶMѹyvQ]q,D΃M6 t A6dmcE/yձwM8E<9n0ey򬎅:ˡ/xDd]vtY):0g!ƠR;vF; >xPʌCp}' jR;vˡ/ w]_~9>xߜϗe>Tn.RAEnd}odܺIl0vEN y|&=<=\_])%qS=)OA8^hnPf_~6Ӯ;㾈:|$u_nNCw'wajGUD{Yo蟺7|}5\_ȟtn1~o#gl9h2,g,;pْoa]2;;ʖ +|9u%t>zJEϬ: SQoE-OxN,'M[Y91E y&zr)Lddc}zJ4+mxr#<|oywT! ,z_,2.+~q _f.1@2f EһSRyjZ/%o\ɮaWx r0\{qn\DWS#G(imTv.?Iq9%`/tkYѺ۪Lw-}CFC|Ue?gH'mȯJ`c]H434UyL7z-r%#u,hY}31!bꃆ5̸]LO)m˕{٢餾Սh((LIQ3htAoG&\b03ܖ J>4o}x66p$uō@);|u0\;{r./-IXޒ/wmcqgq:ȸ#URAyd~MI PgJbh(fM]*VMd|'^ϵ}7ͽ䅂I=νCU{5d&Cd̤z}1mHW?ە.U ((jŽsjz{mv3cw\ MY yV% C];Ȗ|cUԜ6;kAmfW9te('v" xU``x.RΨ=Rbϗ;`: tHg/Qfr$$2y={ѳUEЪKe;~Y]^]l0M-l!|ކh{GI@]F5DIߋAtfk5][q體#ɛ<4ϳ"e>obƅxX9g%]&˘BK-CZ@E->AdrV3Kbr|㭨kk k&`&< QtHhS9rM~ jd\<5E-э~Ndޝ!$ܵ;﹓F /_?N7BӴ0Mhb̫q(;&N\@NRbOʒʹO&sYi{Cju%-IIdz*NC/8d C.(&:QKE:![V̦;k`X)g ' ^> wh7R];ruf"=#DA_08~_H 6ye+[P%U^Yf/J(!Y)IŸ?ɬ>TR$I3yq*c\ˬM?|rzq^.M?68I`V=ض=9E'u.bmGvV'kto(4V_?l0-#:ЭwnwXp чeDzIC}9 $ʉAUTʌy,G?{OƑ_!ewgSR݇}؞a;`7PĈ"rlU7I5/f DVW׻zqW^%3˜n(f*I;UD%q'\A> ^qSI%] cy~ӹ8lm-e*:O<㼦Csge}H ɹh:S?^wϺ&{CF8͑ךZ Vt`xJ% ߼ qظbHpRVY**DSI;#%xwj \g9q6Q=QU2,𞨘|0\AbǍP ڎ`'~#Tn# Γ p-/Wp+RZX" q r@sTup"`%I)טa#4r# i\%gӶ ,4B{|aRYާ`ŵ&WDoRkpY(<[UTT̏xi"FBʼ1xt@5 %a0 K!ps؊R_<Ӽcϥ ~u1 K^^{Ujar`f_dF_:XSYaN@.(BZG@ck JI !3F& t*!4A$Ld kWlNlRzN,iQ+ACyV ΋$)b0=,9 C`#u&X)#2O{f9R5A*j_tVP. ~-@7sTFbpࣇ(6&p@#eyD7E8jU p]6 0aM~$0-mT0R"z*sܨ8nTlW}FsFKȡ:&:0ĆX@1 %:uMO0*rEQtY^-O[m(Mr|, _y0Ss^ܙx;x~t;=w3ޛ^F{RǢaBnY ]M٣gB;Pw0ᕝ DYO߿L`;UNVӀ$\I~}wϊrsrޝoݧ-3\XE@4 s8U$9PSWIeur-g-}+λ3=P1.1X(0v9[<]U%4ee>QO ?RLGJ8($A62KmCRe^'sJVg#FX|ۈ9wN|@5? ˉ0*OUY%lll<-ڪ)ۼY.(~ق?եNi֐SN}/d0f :/.X%#ıJ,Yot?|o.~?@P<.R|.׽({ ONN!>.2,mc'.=SEA6|(rA!jF&1e X"'ȞpU҆&Ofb}?wLn$LVӉ"}˧;ZR#ur(BIe2Er;(6 ,N4Ώ1&|'l[rN\[3cŮHhldG߂lw56'yPasqxelg4V:4. |}{knSqDwӨm,gߝo ϛ>-;nNB mHJ(t+Y 4O[ʇ~~vYГC0j>RMtwwmy6&zF LjJA@ v9 10bާy^G}`-ml2Izm ӤDB*q;ɛ B fb[1jc!̃JԹdX  X YRi1A9B/CjgV7>3B W.ZClr,@]( rk#UXRL=OaX.(x[gcLcl6a#Nc\&00d"ېB)FYP!TG*1L"YJGQO&G'Mri Xh L{1xUwLt٨_ ` ّy~/72E$Z;[͜0bX%Yoq\L| Y;u7 f;z LgR:·;d!p!Լ k+(sL&ÅVRH-+9֟s s*̕Ԭ qCɟ8B:pD lb!y5(I֌:BbB`u=@۶CEKL23\,^BC!bG$6 %jDFG%ДskY6aheϽ:/F@m Q@{Uye X"3|\0 z9p*⥊)A\U9U-,sR-=6f#6Ic8\Ɔ%M:M?qW 3L襉2)e-V\Q,UI`~RA졬F˩܍,c$Y Z`+# QXL|Ab1 rT<X@ {M@۳dzei2'C16V/kȉb!ܑ -g&ߨhI RR"Q bN`mMM4 /m9u;|"SǖKpO%Ӭiri>U}l oQ ׻JcK<^hUiL~G04AT7%1v^5͹r /ɇ,xx|{ڼiwúC}~@P\NqM PrM_^-r-Á49!^#18rd]92pu6lUv:LWL(A5Կ@,kSkiũ/@IZ.{N?!SN)е7!Fv- o!8礐^A׫-Z'R0'[Gc@3 v@\zf(,֙t&!Y[8zw 9Ku%ߠ6o !|5fgrH3dZj8>-W~dnzT7JWeW/ 5Iog ~jR?m x7Zc\CfE(h6}Q r`ʓIi_; pZk/r! }6= vS% bb1hbZZ_~ò~ kڽ)ߎr?bpj3]d)(aub-?t91!cda0bfu˞:HdJPnGNiK8VRF%Dճ95sPu0ϱ?TPTn;.ΎoPPfakJLkުIT.vn6|9 < խ%r.rPJ%y?eIej| =O(@HQ0Pbc?TL~ 痖ט+/\7FΣ&MZExI61zoQ[LK7kh}geޫ=]ukUݕ 2regѶ v|<NoOxdՄ-mڒ{ Oxb+=h,M/4;mmRQبm-aВ'mmnX렴I &+dBP"zȆJΓCMF <*1sCԓWj*Ťe(%I#xDX v_eufߨ@WRx:;Q&MG)SE+/Ns*@Z͓e!R&[$Q&>DFhntT@35@R[iՍ A(Q0h6IDHVs ڱgz8_! R <`0N& 1E2$%1n.MIJ86{RUe-%Uw=߸]+ݶLݭBCZno 9BF;' /h='•;ˆ ~)}n$ E*+N5md1\ʪ,j(<C. .+W)ʭH)dUalf6ʭ1 fJ$,fNRq B3\YaCroL8 `Zy0ò1[O9w)=y )f%D԰Xf8an -loڔF/avSM&FLߢlU;k%-R2^H@(c E<]*c0E羳1 b %tz&{>];nV)UP-`4A:rb&sms,LѬ%0Z7FjrliLy T]IvMVڢBV K$aנ˭^Ƿ~b=~bwi:7/jv'e$ԃú[BJM{мUbzMNոUa{(|&6EK6&rOjYlޥŎa#!zkHZUg$/N) z~QS?SB!12q&( q ꊼ[e(>v;S0}4=}S/D7|}A(Eqnx`!+OcC_[cA"%>PqH SVߗкG9稅A m~:UDݐ`j5`1NbhBUb|5h5i-:E.`12 +&iihQb][V)[rR$wg1!}M" x„kQqRV BDbL==6;;cׁ v8) uK/8Z"B[i"1S |sc6 B +50] y^^03/?ѱ:Kŏ.N=h0/@o܆>0Gh|_a^`ǽaƙLY; 5 ;[Si$ C1h5M}v^ۭP;tb0Ǚ^л/,i|OsFJk9;VεSytvn_lRXHIUg V<lnMRSQPZRPKHPtCXEx*)k)m0[q|'P)"r,rQ 9*Zd+ E(,p~4nc,)@,ݶ`~kk$ˑ%X"`I,3R&p1˃0x]8"h ȝƑƒS#adHMIi`->8 l&V"_/gv&EX e. GݨMVhɸ͗L4_мRI"_`B;߫y`C-eaoݻS)^9Z_nîV֗Z| *'G  NJQnî v2$ZP->|G2;L2tvMdH!JU>Wةb 4Ȑj'C ɏ;i\ Zo= > I˳QXEvfwlٵDv>0鞝atJ0ShblIzߙQg‘it\$êԵZ5 P %1sQl$2gxe#!~fo{93^υ ? C&Ԝ$ŧDo7ev}=7եr Bg7W^= wޕk`ɛX^/;7(!_{T1ZfLEHS#DF52Y \Äpz4adMuw=>8ab}qRMevjЖm&Cuc v8;\ǡ4n'C *p2T;)G*x㞝d֏[]&E]3ܐ,f2FrC|=R˲=c]㏋xBvU؄Ol2 4攠xCEŕpO$cpQ"vMFDnD$FZ|qKM.7[i5!Nh+ {K;2>PvMd!JjGF;]*-iWdv[K-Wťv oHR_|=zh+&)˴[Vw1O;ؿ6f35L'<7_L~s.؟& sZiUZW-/*]JS!8_R?e8"?O:gfgf~btV/4\{wp7Оw X≯N^ ,_3CPx-vҝztje7]^{ q lv]ď'·pË7TV[ |G/`R3uT8ysfWÑ9/^~`w4l ܎kK{eՔ/[5_4r0O{-j^E{̃Ê8^r_l<ُ/zK@:;u6]^"Fj/ݧO >7TlN|}fF&1Jקso/_^@{WO؎I٫/`{B^5S?;Ke]i?ԋ5K*OXX?[8Ӥ&^Лat4z:羝Ef,\ Ɓ?AJ"zS2Hq{Ag\\i\|y5XT9fsʊ⪞/5V|rv_ѳlLֹ'E|7;gFYn5˸(?qZpcs$a9z%q]tI1Ra-щ~Ig7l > rovɷOc8_lbsOnNr#,)Ni1Vf\sfI];H瓏Tm1ڸ]sDsBBNm4b1[dv }d킐eP1Lu T\[n|s青1-"%vњ]bAm׉mq?4k`bM :>w>^8U=N+|?ςʧ1[TV-rI[(ȍ*bK8 N1 et{f̱ .hs%ΟyYv&C:»$@R=@ ђwrdhpJ,IP_bN%0xi.P*e%7HLCAL}:IaH4_g?fT3y,'gepC9iV MS@'&sC 3>As VU_^==dCZs5ZG01ͭr&tZVyEExdն,(2m|Ȉ2{mrsōpUyc"J.Ps4LFud"u`ɥ|O9#{$QjBW{u$腖|{C^җ I[<4r%p.]˦,ZҊKrg3C|vXi<,HG[ӏ]q cC 4! 1±ё3J?O(p7PVUi>t=&PKMČTwUk &:M@i@Fۯ*Vrg9g aEB9B3tj܄boK8f/xN ;o6C{ ~t)n]&XSawۼG;T|l?XAP8;[wBZ* U* J_IH3Q,Iec""TV!6ل9ArJ^<3XlcI('GC AҏǡeD݈ J> *ܭcrh!'~i93/, cGzۃA.ukLE_nNX2OR`T@D7b/Cb? e"IBvI]\")[PTI:a F{Vˋ#JݎT?莴 C+?1+?w:בuߡE0ahqEȫ3?JphEgCA^3)14`Ȃ2^{Orȍ]6rD^@tXr& уZv׌q(/Rs^ 1aY",]u)#yJ  +P{xEA^DI"USRZ>E*tw[ğdшHG,.Y"eSnIbmtj2꘶/q D2<`M0-l%?GKxmD5cpjsgϺ:#wb_|=gsWRFa>,<=nukk%Z![pD$R HAOH S#DXOiÓj:Q%;59(I-'FEZ-5<8fp bh[qEXٱR읢:Rq77]4-B`oD))%XNfn0,3i^fj>Y7`rmr$Hb$ k4jN[Op݂Xv T!I0-_A-&n?DU*74U86Sc)R~)&IdcIm![nS@7iNi յ=4y&T}o䙨I.kej^gՉkހ!26L_l+ E۷҃o{n[:DكS)V\ro䪎GvI(a*WUg!{%#X镸(1,Nda2J[kYJ\)ﭡ|?6+QU^di<5X|Ѣ4rֽGt_`r"Ӄ-J|ʃ"  2u{]Xh벍Sgn} ֌^"t^.Q ?SԦ}{y BZ'Ɖqoho(Z).bP F  FXHZshTZvRVh_J| Ec, :"o˔c cǏN$]KtM/EIW\ɱPJI")т9 CFVwsa)PƠbƆqTj%|;x sĨ;<yc=0FEc]'d nE͍yI) Q(mpgŢGqz˼f(nq^^xL?m;רW{~7% P.8V|1 A%W|a"xq Q0+G,SYV_6DY@)6[Tԇ `QLO3[^7ET5?,懃``mnw df毳=kYbCM6n2[ *2Wm\Q~Rp,!gش˭WE.o]1}bk<_;n@0hi4c"N)!yf3gȈ+ޯ%^D1zht }b'~n}=GL?Nhr[8{RH.[i";$[1R*cM`>`}JX bd_rR(vO,NF= O "Y8u^g%vN$UK'f`Uw4J^ a=V䒁I]g@. Hji5ue߳9 -qSӱγUnKviMʺ~ln0[ac0=`a=tBy% ,G7쵍u }nzT1:ه=u׶I@ޟIocD:a?MCeK XugBzdD /[Ö\xNw=h=4"c5qP=v//Ɠsoξw㷇7wgwZny7EU¿>|}r;hL>rZ}w@ =oݴߞ`s,?6?y~~kb~+~H70 dFjw5{3*٣n\gMjO~rgMM&`)-2"n_gGsbQEt|=oҚb7d#`uOp[q-^/C_va7xd|Z ig14~*~$-k_=ouF_<9TΓ/!%v_2jL0 }:I4j`M;/9wezYsH}RӠ8'voz.mq{ҿϾE,}~77W>V մ۱tw/:sq^hkl@93h}tY\oe|:JBf8v(J w`8i(MA[sLQ;f v%`. UBgM_(n|@s @/6gVp:y3N vObxj|{Hm}Ō.,b=]I %u$B?TX)ӘZgX keeU]),9Ky`Qb$O%ԤBR"S`%#jJ0OqR9(:*EVRSBCZ2 sMƈ N1}ٗ;$870VLbΒLiA9JF)^Z%Fjܙkk?k_oպW%e}Oβ QHxӄiasE}D q<hr (.cw M#xxe10F;8NEL985%Xr#@>3f3>tE/{$RV}xxV/cMZ!s۷[{+4|fĎ/5hzEC#Ey"u-y[*RƕtR4"W& Vۛ֍7ozu`Qw|x e͗uTk5`k;ګQW-# ]4= <':/qE`ڠ3be3ђͳ}y! 5Bp褙cY Yon<34.)$!BQ]y4'@\);R5 VlxaJü^L1MO.J(t黆9ReiGVXHؽқ +M~%s*B1QY] K~Nث:a:aNثN'|)ѿ9_7)FQ%,s8ш!A'!ob wuO\71Y7K8gכxHI 6xTڱP&e 2Jp|J ݐV:r#ܿaÏ\ȕVG^H?5ju)Q*1h>HSP0Eqɒ d&P2LSb~P 2*]%%_JtJ$#?{kTIت9" L\BpZ<2URR؅+U2,3㝙&.с>W}V$ujHr΍XHց=Gネ!UHA⏁w;H'_AT"-$WF)MOj)>zSh}y0FTK! B CFu}rBV-pa+sYң8ta+mApz8N/8USU Ty8\O|IE4b\Ǒњ~6ڄeZTH1O: Eڐ@ڀ>+;p?,$X n ̊lS?zfJi:6f!fd+-BTs[@]2(=ĢNZ#wOJyõf!܂6^7 gA^_ٽI{&']>r\T Bfѳ;C32Fn_o {Vm +rĭoE.ַo~/?io/VE]~'+8nWYEyspK=a0,F4;.:`u,#,HƓ%#(!huit‹( /U )x$sxv}co#]M)wEKtL%Wv2~`,WvF2]l F {:*kBKd݌̔wEJg(I u Wvhɢ)gfCJn}V B)(7ʄLR[g0Ңвِ Vgt`$Y ?+At( ZF;`JDcc[:])()LI!*E/3X'șs25YUVb0jȻ-5!5Fx5RLl7{)iQs B*PIb:Ggުd0;͡cX"TU*u!%YJ)d >3o#Bd@k& ,s2TR6}ġ<$K>QgR Py^y @ª(.i ]L|) ']&[oK];d"cje9)7(g wrfD!*q[ZZ&.bXPdЌBzKBo= !M+0q -2D>DGRNRj4+H>Q:[cH2Y yfI$ :Ifr̥#2z,2fɆQ#@ьӆuUYED:2Ԣ7YLjd%Cv7 5QJ0hI# /gmQ'Q55jTTE;ޥ6ѴY-g|C$nA84-+M5*Xg)N1`lޣ涷ݤsD]J T[/r=ϘKAv.1_\۹0h޷q;vGwI Ō[@9R\Z;ض5R| zj;X6œ/3)zviFb[paL0\G@IKOGTB* hTRƶ4 xFR>rmhL3gR4)),y*ʜ9ydalv>n֬h_cX"&XLl,]LR砃yRȩDT$PEÇ;Tt9YH x&&8$5 DS#c p?FYW ;wBT2k~%RqS٭v&OWm+Z ^ /R Qte[yVF8Lvݤ@gx~bL3uDNKdTYd p~6^~}-l3Ih֨Q3D;Qd$l=%MeOFFGeXf#r5 )9ZdbAF[72l)B M͙[#5gIRsm/R<@$DKJJjiYwtNJ#Z5=iAAD/(_cNњ !Z"bS!-|zN~;) ]fJQ> Em i 3} ۖw] P>QZ^\ ROSev5*ː}hR)1q  =z;C\x4hi8jw1b=>ۖy PuĶ坺,֨%@'Tu%ŭKgaفĊNj(.ΚsCc('KϚ./?9E},}MhLz坆n0yk^Fq֜OZ@NL*⩑A ;x\J1; {N!-]F/9@!S BZ "[v`H!٣<${L1$鬋fqbƙL!s杤%MZ1s}6@K(<E-)j% EQVB,M)ź{2{ɝȓx[L4}'0f@q |<{LX5/ڛ((#GIɹqiRQ-iA4}gIMTP0aEQhpKz_{¶坢BBtpݣstǚWE]V7C}<*Uo_]<\W|z*F'U5,8]/xu_=]VhZƛ]txtuovN[,dǮ$='Xm\ L*X3f˛ItVDO?ףb#ܪ g0۝6Ƃ~׌O }|Pن:I/?xD je`|̓4l$iŽ\Z&>ѰJ-U T'Z;^ϙ,HQFzRv%"rP+XNj% ֢ݭ~Qr*(#Z7jnJ7_.{JsX?nv=51n}Դ Z\Fzq408ڴaxLjr tYfZ|V5{@]!Ɨ^=co@jC-gV34ӡ:i ^[fX>r-z ~g/Jl6AO  l48UQ [1sEA^NJJxڋM>QV]Eo;do`)Y3Jd^|RQ JvU1C+I[&TKwp7tkIihd=qн#BI0̩͜[͜BC^qHO[r̍3Q%~hhY˛OCKԭϽMXYuzx_;~ՓćYc<ѓ7";hwmG~ztۇ1Hk~kԙͿ-Fgq{HͬFS}KzF_O H#I8Pq]c博E\7(sLcl|.>jO7|n}`𜢬NDYg _N0sH맫#nxk/E\}25hx9??}?xKls~W7ثjy"ci$㵪o/%Wyr#'iE kyVS#FX6>8ZBuitu!'%{A`9~ԃvMR;bc믟'j&e>7:[Ͻ;1F k _~r$yOs; A+qx"V{V0aXN9ɟ~JWϥ.Ԑo?~K_|Z&FrJ44(%HQ2-4Գ9_[EJӇ#D -SϔhѓΕh1brZW8;")NjB ؐFcאPh4$%sLIf֥<锬ݗZ&D-S\4l-&T ,e=!N-hfj#e;UVary?*OyT3`vYA2Ib`JBBVhlJt.x *`1!LI8pS"Fdn VО-Srf]AY&BJdTzلMR!W R%jH֦kFYɖ!BpBHe2dVĺ"0lX-i[Y\/S'w7# I yl,YXz× J^:*+@`iUL 8GҀ| xe[h^m*H6*L4;'dQLLҴeω*bi"9is1(դDݧRƽ;Ah:U\qd/"bzGLDۥP3|kծA;Xը\k?#y[5 xȲ:(|0`QɳBɤojl/ڕ$+eoi]\-֛Vtwݷ7?p{_6?nEs!ݗO~J?$ ޼^pB9gsihJ}SnZre}8OwVW^]ބw>M[}6X ?.1Abs Kgx2:޼b >CJX7@^' .anϿl'Q\i 9j"Fl255%KaQy-wc]m" k B݁w2Āc9.JzY:fgvK/ \KƋ%PRRȲ(O3ß̧HSDŽP1z?4_t:r͔ٟRq>5^lX(({c J@mkuBE+& 8p^E_*%RX ˤ"Y`CaZI`XkrL0M0`uue%LTo`\w,[ nf}pe%C zR4yףɕv3Vd}_f][iLЩalh0Vz|ܞi߉XЛh2Ll~a'6X0:G.| (s|:k$Eė#9ܡ> Fˍ.?7q#7q=G.wjj yz-jKE)L(>^Ӣ_j5K-{4pRpKBK]i]}p.FSR+pT"Omz8._jyQ&woMó@$?$˭E}"!jDmK\zl2<_]ooR|][ f`^eiNX*`MWMPZv|RXrcNy2l!O-ud?{On,J*6638QxFE=Zy ޏSbX=Hq-N$dIƕj:~bE @by vw׸wW;a*K1BVBԮc"fEw_!Rl'Y#CmJn]m)zj/[tV4WK.R^6+rr{rzÂu֚uD#& FO%Dtc6 8yQKIԪ#%0:7ݶQ:D4m#%9k%Y,@vg͙R|Y׮B|Y.ʹ&?f];F%ZX@zD:aF01@D`ƐDI6qn̚(gXu HI1r Ug2 CoP;!ﹲm-9;:Go8S h>tYx66[D4AK^~b*cF1~itv2!A7B0=A GgCmBo3 L+s y"ZGuXhkj7ؕNAŠv;yjZV5!!O\Dk'j7)hcTn3#O$<4jZV5!!O\Dk'Xn˓nu1(#:uǨݎjZV5!!O\DkTҩ8WRvAѩ;Fvĸ5QPڭ6e[ELQvBVvAѩ;Fv[[&vkBB.ɔ*'TQ\|]I5?\d#~ f>?+)SDŽY1Pj#׬"9Ӌ$zgx+(kPFƦF0Q!t7Q; e?s}yq4]6~~⯭pu_߁4vݯ߼LG}C{]ώ֎&^SE}%D ;`TD2+aafXPqU\G Ky|1lxql]{ SKQ角}|t5]~~,~`LgaT}R컀 bs+Buk!xOmǓ}܍@hreݔxgEO~Q~.>uhfޥQ#['ug/c2u +A ~Ȭ7;egB׼ޚ^ho*=POec\OwF3MfUBJRHO[J![J"ndHHTXTg3.xabobz>yk~'1Ϧ+FKt/N-3]"ꮜX}"BЁqfBWьOMVMI(*u!NC椧Qa}JD\^i͉5.VϲW~~b~ Ҝ]Sz LeM>{<8s]m׼yͶKF?nvzC0ȿ$ޤ;'-OU7F^aw^>V~g," b5vGoӫd&sVޓ>boO¾ m 57`lxw^ x~q,,8#6n"5 s ]uI/X,c~V\.e<s">㛏;2;V?/E+.$ۭ1+!\u+BXr`[@BѬ&0YI0YjX=fɬ6'œmiK{Tbâ*xU Rv"1'#Vhs2ڜ6'8ɜ R h3-T@i:Ϸ_ D6".' /6|ymˆj o!a9_€QF  $QA$F% ᥷"$Z)Ba\bsfogXT{+۪qw,T{uf9nMUءZ(,[Fth:pYƬ`Ab藄<ػ".I籠2a;(9SXMJ0|`J*J\ Dt;MhZdz`C:OY6 g5R.N`x`DVxBUB Lr8"2mAY|JלumsƷYXg G݊PZM~Hď=B $jIXa-$h1&c<./jʹ5$0aRT$bƨQ) RiT좥c;H7XT,#wAX-83R9#OF~vŚ,ٟնM~jEr O_$QG0"\Yhͯ/.?՗*Z^U&74WV\ryՠHBNZC:["iĩ~n:^v3}hQ@.GuGx>[&=,!q./2ZwxCQm#KAOE@?,lO$6일L")5ݒ&߷dK%Hjl֦h;UNIf͡Qj[mtyǂؗŧ~>͝=Z!eE^S4~5u!j nGU_(4]G"a!0:KJtg2D/[Qvu-?l>ONJ.^8Bh#JYD$BK̕u8hfu^V֎6Dq_`B- 9AdUFGZp)3ѝZL-~LĻm45C s\z;Ųaj#^mGzQA璷NJ$k#ݚ 4Dhm2[.Zo][oN;Di#x.'%EWء#"E`m\4Ir/'Az44"aIZ,Q_($p+Y3+^}勬;h@7f>&}._Ij>TfS0KK1X񼤠W5 H$/šޔ[ ^w݇ NSFg܂B;SxV*^5V]H(/L+iY-iJ 0FG];uX #2qrUf[= -"&*`K*VAh\ߡ) qnT'RM7z Cwf(n݋*`Iyg)Du=JLA.pg`RV|Ewl  jt*ď/lMK7Mt'A6y\jGHsRډS2.])N꿆E_ '1{)M-CQ7 i['^ч3 FPH:nb3KCQxE`4aw <y/J`+ ShjT9Y@œj3.(kc7s r$t@rR@PZ/q;ܐ UYl=I6`D:{`dGRͷJJJ)eƅ7Onzy_ @: ^9h֏gʎ,4PkH,kP*R::bT(UPl!Ĭ狻$u+iOˉX3DD6=\kQXBz@\&TKGK6o^] rNh1{N@b0ZxE~vw@gENGs#ui ޝYh]\J'.),|xe\YK8cͽ7ߩ̰H2̧L$h&#PP@>?  y8O/` {J'svGOb1RVKʢͶqDmS8i w((S%eGgSF/3NM ㍒tpV2T(-( ̝gg-TO ݺ[/6xejKȹFZ3;u DӽY8 wR[LG> fK:%-cMaS*5 #jQ"Ѡ - #(*D4X?҉(M 0 bJRS 4Q.bp݅Y, ջZq?KMszR䨪6d2菸/w;גp3c09=. 53C P S6WނZÐu3aGd&X;.fUe҆Bp]+>E*Xtmhy=Y `)'d_~1P B^$ q̾w'PBvc]AZgJW-1,Eev'GW:*h>^x߄(6.jHɮy҂B(@Yw 1q4ƫ۟-աbĹu ruU_Kn"ܲ'rAlVfq[Paim43JZ" vuS!ѹwi]lt- ={~'}.&U%{h0D "!wdб 9d%`B֋XZvmP O̪KoMl֝9qZ+Ȓ 6G,Nyl\7f#G)X=w'BY.)X>=#W0ƻЂ NEC_>~<*CݎuQdBUѪLVqL  U[ѳ+S4t }!C L01BB2Ճo;QJxHC D4d!*f0E!l(!Ng bw~ƹ ˆt c>dsQi.-C}Q'LpCn64K b!(V(lΊC,s%i 2H4 L<iwOgJGEz|υ@Mcwc= v@HI @Ѥj5*2;`螼@`(J\vpc;'$׾ࢰMgⷒ͑#=q] ۡ<\-`LjnGc3-v>͖wݡr%Yo}Y}{rroAa5 9Jp8Ԧ݁R%5l]J G)0Ԥ=:P*JT;3J-KٱSvcD@!aa؄aZCHC!*Nirč"E2/1=^ٝB&GUdn7~z脆S/_.Eo ߵZaM/4DGir!:SF{ 8 ~_d~T4QlR,6hTC/H,>Vmya>ퟫaL7ꍝI7xSn.mw?K|4z0|P¨E}k\LntĥC^qslxey*]/hUjecz]e PBڟg*Pu!Z6ȥ+W?LF?V?idpO8qT~Oqب7X{録W馠={/ {zgwJ{cMUvlZ[c>1 /HѸ(fud5i+}wEځQ}z2KLep:] \ȗ㤠>zq0}Ën.Gv`3.ROm5F 8qP1QT>?jmV-JmYg_[%0eڽЅzh;RʥSI3p7'tM}TSv1P|{rV:E923k8kg-x:l&\^"\?+]֬ll~U*82Wy7U1w_R ,l񪮖SV v5;\@#CF-#r+-PNGx$ƞ=nQ:QJG~yXCo{j˓$t \ y<&DO{Jwؔ^Wz$ZzH/2`"Kg _></L?V6r]aa^ɭ&: s&dPIH%B`19OH;$&G13%Eq:L~ɅRIFsG)"y2r 4Yn sZreQRh"I0V_i{ x٪v ZRw+YLVf?:ٌH3I@*퓜ڄmQ˴`BUMr‘6^62c,v'Em,'I;p3Գ[eۣ׸n8 Ljբߋ|jZ #r;Eb}l=˟)E_/qǘm_ʳjƘ0k*v6zin7nVa})DOi<+IeQe3.J%uc¨?p)/)Ws 9 Fl)ЁVN'pZ^E:nZVlZ'vd%J_b˗1ZDe_:2,94A |\'I␤\Epms@b  =bքneO)LݫPi % +s朐"J_0a!LJd2 s'$8,sΩsfrB0^uN`EΝuJCFVqjXZDD jM54X!8^9Ilf鸷 m&K҂dz)2;kADl$#h);1)mF&\DcdJ <=zOz[,BD'v6mK|/^t;JLڐmkCWcm㱬p?o~ 1&,{cy1N/aҼ 4LAB|]cNV`SpN=ZzAʡa)k d8̜ؑI2.,e[ΧfjbSVvV);ۜyb,Zg=ܘ\ sQZ%D76}p][f]Íwg0 Aa  NMD=V/>^1VDLmIi4HS߻5S߫_g( ڍOOv+5mfݩXXkWU+S>`Y==81$4-ZG_vhʝoɂZEthT K_1PpjҨK@IF,ZD2F-T"2朄)Ѱ +1ƵNP&hr-cPI_dOAIst 'ۮ"'YD3!f/")N3%d9F;CPXs  HEډrT{= 1L[SlC N~{;k?B:X,Brmbh6"2\kBBrZWV-z>kI)Gj gΦ NC#n*<YcT[J5M%޼TRHW.2SlC %/۲G&05fLR {N #^yCב@ 쉜G(etM@M6&ԍ0 f>#ev`Pev灍 P T/ 鮬YGba&3;/~wW걨2.ŕp@Y&Ȗ*oͯvae @1tIJY"yX&t&4^?qqé *0:1X:bAd[7/2:73D(텎(HGA:&"voPE!|=<Ҋ8|8)#O1 R*IIO)Ir*$-hhDTT)5EOʼ*A?l81zv1w7~U%n=~n(CΨ2M@ CW+e2VYvJ'R!d@ ƀ0 H>=-ʔ0 : N{ŞPqqp"GCbZs B%$^0hLz ӑENyK Iv arMn7c.ȪEgkC~4G$|R#,W)'lҦLMF?_zF(C>=A zضouuqF{u{ 3`Nyu=osEW`^`:[իSBBLgcz C=ڕ5jpw;Ҽj>@9Ӏja@*]X9Չ`y#xܞyu>`R$բNx$dx|d^rXQW--*Azuv@)`Rr ji$SN ' IH1t@81JNd rI;MMW^ç>r9ZBik(9ba2ay.)Ca CnDR\QG^|v-z7E{j˓uw7'=vB9K@Dl!Mס*!26EK,]ADDIvt6C 3D\D(*s<eŹIsͭqi%W&˯F+(ZxU iS^#AGB=J#iE-(3@k&XXSO-|W3{%iTĤ 4Cغ"m|X+#AQGG]y0_S.oH9_*ߢX8,/>i~mTޜ X`:;>t~2srV7oN0+SΕZgY;;k.?_L G6x'G ˆ!ԧPV X1wl/F<5@m쉍9Fl +;'}Ԍ$[Prci>5h,8߼D|c]Udg~ycæbȉ!v3̲awggѧ͊w%o0 H1f?=,-!~*-e,Ț.kyXҢOSK zvztPQ9(Kp~vdrBj}#)]Qbs|>Jr:Gxmn߀Խ/iJs\ƴf/i]1mjTZj.Ǣ$W|%rKa= }>ՐS1`N5ݝ5'67WJ(|suo96gRIz>+|,Bǎ|WhS̶tW]pJz*Q*̶KAYIW.e B*|VnXR( w`"S푩Lk'8AXBH{Ojqq0D9?/1YST+\q6+U5oNg{s;v}ٻMOUoI8L 9Gk+gPg^pbJAJhoъfeObtJܟ"SSr FdvO h!w,OmصYxO_^v aiAOчU%57A{ppf0 -6nXnHr8V];ثU==81W2 !ȱmX1 U#l$?ĉaa>oټ7PA~'P^ yr?>!㦡ɴW&2%Kx KxL^ҏ _YUD|nwY>_M:srgWs}[g뱻 qO[D5S1k[FuyzN.s3V'3HA[|qf,wDk۽$?zr]C+uUXMpQf+XH=DC`?^f$>w[͵WӲ2%HCks7x@A.>ױ T3 2'c7DHs¸Ԙf BB9+Ga l5vW֬y8.Oՙo<^[]x"g:Mӟ8ObERP:Q3g 75R|ĥobMo}x0C>?]L 98=mx@/U̮C TMR  .A:RH1V–}B.bL:c WZ]ŭix1=8ܺcd-`q))A³ `4v1w7~Uwv"%Se`][o#7+ ,Yŀf&f7xs2Xl9+%K;)/#w&v7*X]/NY\q3NpZy=8pW56q,x$jNĮׂr}-S !e,ԊTYjƊQȕ=ErSAAȦ¥4zл: @GkQн~*H1[8dvi5쵀q{.j<77J.Pn7Gb^ VKWg_ U% 3!r>0qw]x@ݙBՄP7: N;j4a` @^h5E.Ұ*"-fCec1)K 3c WZB-Pd0}Yʈ**a<1F߀}C?XBCyemdRHHgRZEj1@٨#2!Ӂ,w`ZUd'BaQLJ .=vEc]웂@P`ΰu*- -+P o\F[Ij1uc=yxU$z>_fc})>'٘ΕWmX*&Ru/e"m/j}<,Ϸ77!୉oT2Y<|q2s}ŭh3d^:YN<-tqdQQcRLV;n ^GTv Mr y81C8I'Y*ӁT! &{=y5eX!3hm,!:gM7>Ntf/O =}֘-YLrrx%gfH?fpnkշ臻ݭ=㩱$Rw\ z߁d:IA(׮EYm\[4iK$IAl%GtciqV&zuhꔊ7_v8xn;/P,# :QTٕ?_w v%ח۶)P YUi33?bJ%9C5y&(S%fKؕhWtVSfˇ݋Kx Қ.Eնvf>wC,%oh a.JrxtFeG7W+.sqo}|-y.s'(9.Z(w ;t ;a7V ĕh^ṕֲQ;_t2e[4oh=N]vQBtnY]7?,ʼ"!fo$-a$P{V%Kh!RJ̛I\fw#`\2b}W7)DZ (.uQS- 4/OZQ<<6\Pr@rAxN.٦ B A[oXjrdMNB6}kQr h6h/o*UaM UV0Ƣ]Z9";ϴ~b̙WuRmQ3ɵц>`\h^e]|1ctXz# |_Nrql]V1rۖ0}xI~1??=}YAWBg1Ox: zUl0%@SA[.D+]κ=wosY=PLTʿcw& 1C~~1=V ߕ٣&Q1>F$b 7\l:5Ulo.*3' !9<BlOetF6>u@ELoxG8B(  UJsRD&L1 !?X?*Rj@4Dd wAR @GQs!a}y/ =򌷈- ֽnu20¼)Vk~HWhb,EuaF먙۷w!xRu(+>]i?`%$E5_|JP}It^6fltP2$eyB A, D & VȂJ*PJeH4 - 'atv/O1#蛑*)oPHqiY*r QP?,\1sPq]!"DCVunq S@4nDvg]k{&a&.1CPCC0L`ÁJL֮"%"Yh^kwG~O*8zīQ֜ky\ƫqxU8 T,?X1 9W 8d^RQ\u9 qaȂV\ 3'4ٿp Ɨ.ӵOPXP,V(3 SFax^wզ:kzi֠!Hͣ v_1~WV<¡ӆ GqHbD[k^_ݫŊ@B;P\'bAiwaHBÎ` Pnc $:Rpҝ:>z؁/%a j~Bāc Huh.XU}4}. `$!| "qo5)~@ \4WFnuYb*nATFYAh'LLNb5555ּA6O"(RL.$I)Lq`&"II'r y60i,OV[A}Yd@e23ś|6-gFUyoC4νcҼ`hytڵl1i.pNLP%Bee|LSS?珿_Z=))'ahKti#%א1.`'>YE1v:as%M͍:>ڸ"(jS}Lcm_7 ,z%p-1K(ig8em8nmwtB!nFtQgieHhIFZtPPim^_⭴ 4WVn^?:{Z~؋lܒbARl ^X.n(M:/Ń0c\!^sy8{ѨikÕ8ʗdڿ^`Nvmm.E 9c~?ע~/Msu|T"f&|;)1t ݈Q_Ƅm>\}NB*9)1Yi~?bD+]?!^GU|Izרu.JEAEΨ f4~.gNһZbʾLѪ-NTpSu7=g-3*5]|e&(WLM_&y4]vmO:4}yUˑn󩚗?s|ѝV5?둍rpK5;ux2}7uGK)eբ岰"qJ4yt0SUba +/TZ*wޞvWKZf*}jAuj[!,䍛6xXi6.=kw䕇j wࢳ]T*K0G;]EcUs-ʉ]TI:evjOzyˉ~f*+Lt(ii jjU,oFH m4Q[0¹vW9VwrѠy]wa퍠ù6jmR9^[.nh *>u%;6^ŧ Eљ{NL+1{V(U k4#Ur8VN T"rv!z |%g9bw8B$,1S_x.X kAhSN%'E\2'' '̞_F{uE"zl5DWJBGqh "JG/*PAH70nZ?`J`AmX]Ѻ{/PB,ii#)#ңg[[p̏G8݅G(qhX:hVui YSΞP YS %a9C^rʲF A&ٗ똂(}f! s|NRAAPˤNEƩEXNti"N&v %Zgci1Ind׵zLuafĬjBL} c#$DhXUХG QHa>$.`M:^: I!#v+1HڋkeZ W+ Փ_ʱZ%a8 R,tyۜ?-L^z7I0~RM'q!cۇ͟LlD޹rϊP-T~Y V 2 D|x+7*, A>N&"{( senAJi;%3lԻlnV ^g%GTv  <&y6*`wWN_&%Ff4cA^h@8#sǬNÎ CxL}&Wіf73Ch`_QXdgo&9hidF&ZU xc&sUhҋ$!`I$zPPI0`,]i7+ }7ȶ1>$Xl66dbn=J4#o[#Q60$멋"A$ ~ybr |]-Aݎnu5$^o|p^Fv{ܡ ~)[G}{raOߑ7Or Ek q6DDp%>G';^wa/@6fl/G35]M&ش_ jۧTS@уl;7d<|"BIS[T\Onqum{$(%ؾox.nFG YLk~i$eBi6ؠ R0Zs$N8qU(Mx܀', s1($NxF,bFF'&**cp#K'<ļ2*  2S-Px$ $`[SMx*tfʷx9ZKc)@$$*6Lq Wb@5 0,-XGy8~"j7bu6"k F5޽p. h}T 4nQG.-mQBA=;}}YodW`΍}|=n@J\t(DHěQ> Q!%;:1O0 PHM#7)V%AID8UrKVS.b#z=4-ZK:}'=>TR>7K`}f|g;iX o/=-b=6L׻l[Kcm ?d<=}d#/SH? X5U>K׵Y Yk3vޖxRi_Ak[{pPf1̆8E+\rZzehGs)2n0ާ4"25߫W_gA_YKVȳ.D];O$|@-k`Й&U1݄RjdmubuBOU v\͌U .F[۱X[Z7^,'.),~а* G5 v H?7O]EU1<]uc]i< I37tjY NcTBE^Z};~4OY_ɍ;hcZCgl?_ !q>czxYRbE}yLI?ug(ݏސ)pJ$&Q.KBG Q"AsT Qۣ;ˣx rQ<+HXU8H؋C˚v EK|$A0Ҫ~F! O/1viY&\8_gXK+US^wIKeaP;Y/ 'R:j0R:^D^LOG !yg}e6jZ½Zg= F†~Udo8@9WS#?⠖Ȏ*dG4t%? 032/oV:~젣7G?ʴ\bܒ;*Cg>;j7j8V!${nLم7_=Eg!]}sOp.4˒[B3 0XC7PВ&2N{uspn0\JWa=QHd>xQ=:Wofʨw=R^eig޸[Oi?On uszmwwB.o=R҇;,`?uCx {GgiF,:nNo_dEлftTyMi=-;ץ\qUwe-By6CA(Om:2z·0&-w=Ocsu{ P>EZ=V-cg&-Y0{Z7 hNgX n~5(Yw}iQhfwuvhz݈ʣ7_g}9)#ׄoG31\1?өlJ0, ?8s(HUnJ}?W(v_h5/BIO{=yדy=~A<MC#g`6fRpl V $XGȊDH!Ql\a.?ټOEw>J$t^nGYC,a5.23s6Z>#[|-+1w<_߻vk+Q[g1Ǜm!#:#~DYf=|_4y;*!FSh1?@`0Jl䒄Z81Y$BZ?T!((}6ҟqhﶀhW$*w >V{VsV͈F>ڲ6ց]eu}NSC"ґrL4#wo -K^wߍ. !"B5-z,šޭc|_m/,_&Cgk~L2tXc}:x~0)(,5o sC#TڟD y%DCb?ذl`Ul.X\`ughSahts.oa7&~I8׌e< v[O`պ\d ~8Mhj?kY<7/ޥKRgp** O8`${a +pA &T4u12QI =׃-Vc@b̠ h`rٮ%wAIvWlƔ$ ǣ/9Ic@0(uz6NG0M2qYBS!@H q8AQby,b{d$aXcm8I5 aipvhF3܇[o+82Y ; d}`kKY>-0fم~QjtW/&~ 7Zbʿ8#tCPe,$C<Q>ج($k(%[G>ZZ7EAEsKJ|*DӛQ~T_աP`A*PjvV#z"2P.p3%ĉ%Pb$jM1! y-LWi% G*6B{Aqd2RVD2:p̴5$r2/>6)^FcDsG3-gcF\QLEo|,5^PmacJBqFk݇b9&NÓXDb8JCQWP2(684V1PM8;vdQLWcv.갉6lb#L Bfꈈ[Baū Ӕ,1ΠDå4:!VINєGD<8kl1Md"HEnMr9:Zr6J&u8&=ΆyW @}HB-RJ0r,8PĠ H TC&qbQ„QCK+P:FIs);`v> 'y-sEw\sr [kdX!MXh"BldӢaqNQ'PArԙΑ1 \iE1#L0 W" 1 8v^o4c~u>^M4n9բ|OSezSx璺? ocj !W FYyWJe H:( Q6ҰB1QL 1:Hx5,1$ 99+\vDmQ-#sG?zۿpG֛Os姳=~\Xh'a KX!,%5tS<i.HGsꪨR0 /f6n9E-F򰟗{s?TKH6->^.\%ua]{zFS~KKi:W .EX]X qC[>t5_ E6)YP>2M'MlaR1+τQ! |&6NOf8sW.S[O?=gˏ"۟BsZXs u,D{dW,O%q!b$J' Ѳ%y:HIۮ?^p'o[ˊdY/TYm| 2hqs,Y)"5fXwMœDw]+O30jh6/Tp_U9ZyGCԯ+Mc^]u?䲸k^'35'}r͋$s͋<#muI+I|r=K2kҙҐbKԋ+D}|I: +NQ8/PIPE"ywarFxXɉ2P>%sD3/-Jk >i=& 8_xc'ot*q(M3|[%C)*rD:犪ĭDuH;GF%Zu)/3YuP>Lz0͸xD=_DamFEOI V][o\7+^h,KҶ\0bKb4OV+0$cHV QyqM⊻f!ZNu-fƢiG^)7VMzPUފOM<\U~xʙis*)--0i@qVX ȥv^P*aJ':]MP*Kf@ݗBlh?n~x;`>N+*TlYl$)BR1W8r(3hL$&R׹tsHbD7k%5',312 H#2ȺeH zgw1AŬQ%!FBcGJ_͈@@%]Hh)G FQfSpy:UFMd@\仚&"aG K{i'FaP%mR: hi&uy= vjrZ53Hѳ+Ͳ I $kNZC\T̡\&vߥFa쬩n_U8Jl?wu>5(F/+lQҿw_9zI)jnpcqU~Ybo'D N88s:AL1&#U^02ZdQY :pCm,]7 /2fN(+7v.aodsg(2V&:2d~2q6xznOhLjD\:\ܜGww+bڟ{ys{oqB2xHWoQ,|uKNj@40goh"c&{÷_ظEO.]wDyu armT|a'[#8G/6 ۞vk%Z h:>% ڭʃGhZ.}fBk9p6aʮݚ/ZL j+k&,Tp3^jIu&N4G2Cɕ檥4ָe FZ;H~}r`@$̴dY$džUk:woVj7-<!-Pum(2BMRoI*ZyP9hsQ`ڭZ6BM2r=fYۈڭ5iv3ifBk9p`5#L5CAwBdbĕN8c.OmTQ:gf!UU;|{Y3DEJFp*Src\iCPjMD/Iհj-仁[iʪĹcڿFT)9zٿ/2E=Z h d;oIwA:sn#z$MkfBk9p`Jǹ4D5iv3PZo/wLhv!&L[ F&p,eQD+!Zgi$3T9$uc!F믚/`,xpd͌J`"A.R`"&BQu*& M90#ªމK<:ihT'c۫ԞLh%Yn\ʝ̶:sn#z$SޭfBk9p`J1v;ZyP9h1f6ݚ nC !ڄu__46[Be's!<}< MH(S)-$!dêJ<.ڊ$<٦s$#Rv{Yd8D0E*RN^n\wnZyP9h1,!}{~r3m8Dǔ *G9]<ˣwxTڷoX_|]uW7}q~_M~Q3‹[7 U5# B׵r5# A<*k&9a-tz}C|]/_3B;(WbG5$y>\NjH{qX^\9{˝Gj~=uqQw^;~HfcqGٻfNG#bXs&t#? "lPJ6HlZ"2 KDÊ_M(/+U(&bX:)E{e (;nwOND$%-IX"G)G)Ssxl3iHL‚H4ltgx]B1wf`wf b5S|}2L!ҳ-E:tIưFɌA {A4IPJ`5gl@r6fH9aa$: ZTK ѻhz8jvCR@#G̀Ԍy:q351R'yLЂv',R|ݭ_GK[+Z͹k߿Rљպ?ӧ&˗}xE%MC~7+ h =?7ߴ}:^T<ēnlFx8 !B>xv$)Ј3'=*DPB|q/'J#^N/|[ԛDWmhc֘~Lb4Njy g۹.>=|g9}w͞n}:˳FDD0r'X-Y`d @9xVM7eΚˈV$b{bYwp~"6. ‰ T$k"Sl~uZ1Ǔ?xL%˭Kӷ~( s O Q_AYB-}ǝ_ lH>$ZKF?3ǿ>>gӫ;sr_Od爱?Ϝo>]v#~YವdO}xg) &R→K4b}8YM.q+xmKܐ}>NQgM;V8U h/,Ԩ9gafQD\7& l|ɷRДj)@`Lg+KD0VBKw"E *ȔYrǥ님^XiE?jlOeߜudi}ƄK.l+Cʞǀ! NkHKlV:Bڋ@Fôٝ/\a-uWl)#B{| )t"c/NV1LR +*N@ܐT{&0-eԟuXT8%RXz2T;rqPuSᓣgk\+ D' Dz_mD51YCIRf?k8!@gKXL ܻd!fph6s^yXkx1Q Lt4"%kBsa*%c*C!,MNЊCˌzaІ育1cd '6z#ΐlRg2;XnuDl R? ?\^A`(`Za3e|1IZ0+j$,GKH]􉐶-P4WB]Z+0o43fJAMIpX bv8nJ"hК+-YDy %l`CB!鬷E/rl$S 8$-RA[_ŮNP3_H,׺[?5]|ec*㩈c*A"Mݝ!j5ҡѳ +`|/Z_Q +G,}69AA 3D? }-;`ah'HGY[Ȗ7WSvOBAQctTIz)0YoEQaR&kiąWY@(B`ccvQbj5h*)rt^fB9A(fIph! HNBkF.QVQQvYl*nUQ(* ,N`HxdS2&Qg.l<a+bBK$Uif8G9i3ݨHqG&o$BbR 1`(iᢴضΙWu1,k/.@?_l +2$%@T/˓rGF"ѐbn# Yj]}\؀zZ(aY!/wXlL HQĝ݋dT>zT,KaB$G!t$("Z>t F wgv,G4d\ ;7@tf+%2@KeH3sdy0ş#yJ.l꥿`9q[tp:F!幵63SMUg3K/d^o n 6l)duk"[$wYaIKdUu-N[;乐3Vps .2Fxekp9S(xg-&Z l1v-#FqA3sGR--&\?jԶ-1ob[@H7;~ES#&q4rD.IT2&H@]7lw< ߶,-ci A;NZ➪t~vƠZP3rHZ!`}JA :z55"^m @͛"g@9x`H =k=PҾ@knG7ZEsMt+$@i@$?jnDͫPo:# O^e~7/1Vejr O"ȊTqoU;R+ۍ'{:??M'Ep_4dRoH"c h*2ۯ`66 , #KBSzXʋ[y5,)qp)'];#Zل& od e6lt`L,vT0^g[~~"rKu޲k3{G&זJ j|c17LE{Y'[Hn_i~Y[ ׶K_iah/n _w(tK1'L1hAб5ur%WocV N3Y  w+fDSH7THV6s$ h$'N /F,#"PHa|^P ,ϰ5xؖ21~׌Ҹ٪w6_mrpV=ڼ ?vt;[6~Ob˕N]Ё~>tkCٱ6|Xm%I|iet:u,^=f`;/ʣ; ƹ3UArp ``X-9SYC KOY8֭LCz-`FVֳPSۘ׌Bpy$+. D a,GWqf-{kB6ZH *4}>۷PxaAs(dȝh"EkYg!VE@K*PBZ$/7ye!YWgQ6 VPo5<7wz3 k:w]C@KW&7y4Y,_7̤rP2Lg;^w۽֤raW qikN<~TJ#+IӌFH6z?MWXL>:104l 'ZQpTe8ݜj8cZXxUXXq9_CTbV{&3L(T9XcVa"0I1菧!,ZLm>mrVl.NRd:0F"U@t h*Un)nCHiZ.U4$TGX28|8@G"n{d$!|`:"N(gmSt!$7rtOH@=qSɈn1IVÎo0#u -乡A.Z*tȵZHhQB@8s9v̀4MY4qn6}^TP@x&At9Nsa6PWHi'ν֢V)RdKo ִ'!R%+ ꖰ.WC%4+yW~ggߩSQ8Z%TA{g*Nf[5,V2D坖KZg (#PզQި_/?ozb1h)e<^ <89iL{3 Cu/rJsܵdIPReXSR!qtޠrBQzc,(FG(Pyגnz!C2c"ԲL}`ўLVzͷ5X=!;(| O$K$6}lDE;950[ ĝr!zc/e󎓆zNNYtJ~ۦ e:}I1[ݵڮJXۻ4UU="ęHdړ7IN.I2 zT9[pF|{+ҦIiжZ%nYmT~+~^w>w%=~S^'6)wµZZNn 0SuܹۦwvX_}Y=MŤ] Ӫs&9/QFDR `fql}DFp;d{^bBт<Оa=u t;_I?;wS$|vif-N$0yK`ƲcXj8q!GTp 9s`y΋D0 9.v ZSΚ &eҩvTj]LguUխ*f>9uQ 햧+}t6)UtKnXrlwYy?z2G$T8 ~5 ѻwa鮟m |(5PR<]&mIdELڎ&mS. O@k!N@kz)T1gXgX5Y8^*ڈ cVb8j  [K! Ls:ZMpy!sʽ $,CZA>`Kb,@Õ>@@6(V{mjwƆ3#U{ٗ۬ s>͐]þrMzp\+Ed d`VϵP|{>Ne9sNsFK|_`-`Hn/AiBŜ *_9{pL<&ASgEQ!h|QQ`-k*5m̘|8Ǚy WjtFi:PJqa:E[kP[)p5$T WVqm[Ϗ|MX7H%8k]zU%mՊ =Z:x#I ) M95o }6W4)J)m.%pgl41Lvqx&ul: {r\q-t!.ڪƅ*D_1;5΢‰/4.;"uB{Yz L΃>sq$J J*J<\H+rN83(WX,sptg狈5r6wgV;Pu~5 :YIH'; vyB0 U,4,f!timsb*[S nOVۍ!z(# -ADK =j 2 H[lo 0A>B@'[ wAz vVkHj-mU'JHle'+/DKOW_}ǝH^vRz 쵳l&\vR4@bJi4xsX܇s<}X,r=M:NIQ'9~))ɭ!2wϵ rq"hj-45>pjktHմPɊ؇Ւ}`]}u^{j#qCH{%.$=7XI7Fb𳵋Y.],+.F{v/V|Ŋggk+.V.V.V6՟aLhm :h 6PGQKXB 8g@R^T]XPgl+sv>cX ]lŚoŚoXSuv1gr\.WrVkmGE˙>Jx4Ag10==pzHv#{%9ӽX6uhИ$EV%XEf%_|cADq]]\^%r=m@UA*M߉g =HlC`ICaL0Rk 4,1#`bs 9As)۽+blZwo:~;An4i ^@J#Mm64,X.0ql-:0*11 N =33egCF+OjjVp>Gi|v3yPZ%ںF򶳞Z,{ __$ѵO_=x`+.p7wn{ *X;{AMo;b#+]S4Uŷ~[ɹ8@ӱ҉+ENnm׮(ou?m۽>PFO] &Z@ՉIW}{H1Ƌw/m җRD}('9m&}+pM^6nl{p-ue#E b-]W)t2$ q/w'j %yfSr-)e }mbw>yCԾ"Vlz7]aߑ;ݞ܇_觬g㥮m=PZ]4YW5.8k[uQke(^GU,˿˧1X]>ӫ8.<) 9nV H#θFܟjImfjX[m{Tv47j~w7OPizJ<'o?k,'wFp<}L}#чd.Fkx+eryYe;e*JJlׯ+sv@ T:]ly.XTPۺ:nu}4Wk{ usa0'FtN#,;eb EcˤAI+0[42h$* p4edE` Uٗ,wQbSr%6oi^(̧=+Vj@?QG!'ͩd׭w560Yob6}ԙV+/LUkes"=\.Rܜ0rfmf!I[l`b%z z&j/ { ͯ{LZTgž͂cH 56 un57 \5OZ"rl+o;K?G<\zWeKSRϠ2w V=v;M·˯.d/_fizD'Vc̖M$r'6Ĝ$u;Nϑvdpv-'^i4NFg8_j31{yR6/.b;0h]y+'zx:Qw '}uv J37&ݢnhC°}u6z465k mݐfԈƹ#~I6:TfTC cǟ[xEˁf؞vC?CAq]^8é1Zn+j:$䅋L{[wk7jePbPEt\8EuwߠޚvshUVpu),{v]nT8v^R"})o㶛kG+n5H Q'¬wi7l'PbPEt\8EuxJYٴvshUVpuson@vcnT:Nn5l7VnuH Q'~5~c_I +DDZ[w19Sўڭ y" SX5fhObR9bPIt8Iuf8=ڭ y""Sk7{3uŠqv0^pJnꐐ.N2QcE\gA%c'L;+|3AB^:?jj52Ƶquq[-(cZG|h`ň1^E2HrVcI1֑@ }l~cWH.`1fk٪u$H>;wV 1ga$p2|X1֑=SB1 queO}/,ܑC1Mc}5#S ,kq0_Ȕd1  %F-l?_1c"ṣޑ@sFcu$s1"|wԡؚy1;0F1|"ޑo}/̱@1 Fcu$|F>\H ,@y19$>7ڢKR>\HwacFg 4Gs}#B`FRm _yR""iOZ- Q]K!b|fm꣎~ϖExz>L>γO<>Lo'-O~mm6bb:,^KT},'1W1BI_c:OrH~\VuQc&EaW Iq)6r'V h g/:wyUk` CM55E5jv 2+Ț`enSJ1r2T\C dSLͩ҇̋1GlxR:|>,'LYy!sĭk61fmL yg&LJIa{L ecG8]';ecѡ ̬y! uvfǺqv/Y"LNكD65ӱ-vL@;i8Tˠy¨}Ī(u<y0OL]@Y}lBa 4QzjS-wJ2/vvz;$jBqm4'BIѰY3㢇̈jQNIYaaD#k# ŰbEė>$@p`dxEMk}?US ̜pqVBI7`P9+J<^4 0uU {Q8q佈JX{ 5SM%0iY<+ minq?J<(@Љ714 ]0h 끢H0UThB*8 Am39fǾFaJlwΌn\M`PߘtlH=p9(7Ü5wޚk k*'*(]۬zp%HApBhx7/|<>m() ƳRdCd*R mԩlux: 4Dx,B$il35BM-ɥjڰy5SHMY㙤jݗ83YOL i^&x~i1jrm=Vy|xh 8Peݼ#"$A#7S=róV'ůka..x /[*u4~삱&`=ǡG$Iѣ_ z hJ[m@`\> 65=gy r˓UE?n$<v)^.(~%Q? o{;2Uj,eCſSNpi]~&&1rk8贑1*!D {vFBFϳ3kW1co%ohu FDE^$ǟvx$QD02 104 碇TU 'cZ)x迼ym?w/[$NWknl]0s:ɽ-~^_ˢB!]dқ?pn3@MAMAMAM> $0<: At,7#%$@+cA#)Ϣ7JmZft-2nV{JOKɝ[ni\t.U?ư/'4+`XiV`e> 5,/͇˗[a[ 7~(oq!.X$D_G r$" @\Jۗ”㵬P;鿵P).#.E"BBiC(Si{= 8!W$D3 ~{uكW_@ɐp 8JH領: o.>E/%UJz}^^*%\랜y M߿ XH cl, %C,Q >ơTDEC b) d_(fFU)or#n lʈTbBZ#`06jO3"Dr2|CBnD+4* ÀpLC "@ҀEF( $$*BQCctiZ9C;:, b3 H&4R@1%UABq$b"N5#; 3k@2fG42*$4$V@a"# Bݦ@jb,9j! .45!3H`F88i1 "#͑A!F*e$B$ZJ*$:Th@  %U L(@_LH#ie&BQ`9f@X:a!j(5Q1rPZRAH~5ʟWTMMEz8o+pJp+5΋%nM />f @2$gM7rq3 r/7y?U9tz^ʍs3(U׻~~(抗My~x>^"GM͂6Wdm߀{ۛO,_+_4/y.ߙYZ;5`.ļ_4@ۊ ã~ۻC_BLdj[$1'S^к;S z3?*mL4D+cXR3:A`LSqb1 o'wkay~{t>yf%XFMYJF::Ϫ.jKɱ90z=_AqȈ8K}g`_2)O]ENpt;_Xn'!}}(MQ4Rn(FkÄ"Ip9f RiV+0!cs'ʊAϝ=3ʬ_RaH,ҥdY}3/oWWuʟKq,~2 }6߸S~R@&7nK$:'TK;0%F42=phl T!=f$l]gӯ80jp`xhT4WAՍFkJC*9s7IlHc -&dUw_n]ݰ'v{[.1z:"+&K}ٟooSG3Oξ~tuxvOW_3*D{;C{%$@z7V݄3F]HԝaۗxJ2sE߈;q38zaN1[q<[TMD|)=eאa 6i]+QG!ɣ05kUwyvu~wgGļ]HM(DbB6]_^Pl BC*TKJ,i(rN"uK[ro_i'<P /~f qEa!(vzcoN!`f>t{cwӨX_[d T=.(Ţ2E];PU<͝_^Ǜ/eβQ't_]I?)c~0W *R)Z`b\-7zFP )(*qqP' .q*vZEh{5Ѯ'g>eƥI䁒xz@ ]ۃd618R$l02tILkZ;ԍ5NNNNc6S7+1RAJn~@ HB~Y m*6GaVjiɩenNޚ)DuX*kKȺ)\t7OEkr (mV {^U׋on{;JU`/N(nj^=͗LP$x T}d26X,šqbg^!O˧pUY̯Os22x1j짫braӉ7wq]~_m=jMrIcilLxr>X2g!o@%G8B2cpc7V1di*,˯Ő-(Z 6\f{(n_FG s: rn ֐8t5&#е nw'? /dC+,14%Av\|sWt3pN⬶AO=tFS'`e0R?E9[Ije. n2E(!pgV#-c6CZ6ZR?{S'\xw!oEsxBtw &HTA-);Gv8cJ[~%Dq⩨ѱju1n4Ha9HZ}KJҭ y,ZJN6]+9;ڪ=gwYxkx/J_'#'bBnw.pcr^Poo /(z%lVd K?-x [ a IG> $e<倂c'礛MPn.:q~ȇ(Usm:tW˻υ h~^a8f>2!}rֻ̋hh@K"KB׫/BIO>yR -7v1'z!RʈhPzȳ&k\}{]H_܄:uɎ9v(#e_`24=Bf}u[pN [>e o|o;a\]Z֯k؃YyŮh[9{o?-r+!u:ptǙD3wJ ? !\t+`kO}n(W)5кS{QIEH *6 [>)C'Ip,gjyu.Pܓ DB@ 0(j%*6ǭK|$`YjίӜq nS3:RZgxEGAgtoŁ#F>4)c9BHWFa,tsF'?c,.uRQ jW8: x +0: !{`E"v$"T]]fQwOO`HpoT,Eq% %$pB,p/Zâ>](B*eHg焊CB=E 2$gElN#DBaM5ӎ9N(?Lwaqx9Lg5@9+B:.k`ؕNY-eBcOZ*gwc>XUǫ&g;[u#^|n  [uZ\zg+=y{G匎#1!} <9\)QCu4MC4M㔷w9_=hS ~ݾ6l^)O> m3$`W.(SiۤdJ %su,JR@hgn4%"PAİ?n4o/A .5гb d–gښ۸_apJ)J9/Ra$2$eǻ~CJD 13$:ئFw@}b~a@Gp)|b58w.b1NjT4ѹʕ .'`95Tqxc"#J' (-Մ/Pk~ \?-;JOgO?9.{F$Ȼ+L&ގ}_o=tvt' P_㜌77`8+zS~L_]O3<:o%\"۽(I*U?3ƍmr8[(q'>8BRҖTLû$b )-X*SUBP a=kg&XGN"%@#kO0$L+0f$/sl7 "W]5YQ n87C%N @zM@vQIp[Stb Qz # A Ю0'-qnjSk"%cӗG%43J(`S.ŗB{0 Jpyk,Hr*7};V̓/ ęcOٗۯW6ۂԎ b'`VY@LuEwS" c2sܣͳ6e27)iX9Q*! LUnv5b !}͋/_wQLoy렞FRO(䝋xN<gV *-xZr;aQNuMlAj;/d_L@؁?RJ@yxh\\SOɅp&e a2\V'C5gxByM}y[ Mta3=ϭ G}NÝ964>g.EI9s/[J63Vt3\jw^{ f"sLU>.(n{N=tdI&~9% ]<ɘtJ#:L?#!Lco +.㲫>:QŲq6cL3K2_G3`UȆoΔWkgSc|CqbҊH. TbtAqAQQlhˑctd p@ ?LCԻ72I~3idW/f|l -^)c S3 T+kUĠvI\ot|m| OVi{v#)`KZ;fOR_/1ACK垾ͤx8Q(T*p *%jH\5CR 缽N#Vs9[OujDiwL5pkqqyRے"nT$F%egOwNn4EXQN(G6킪z2{ʷv?( :5jP NLgҖLֆ$l_ѕ(MW4R\IzwK5Yf&줎.epr|R5r,e|wq9tMMM3a8.Ӏgy%Œ3-EG\fY:~Rܤp pe Wyo #>c{25\-OSt 3vpK;, ō_ T- *u}Q넼s:6vZ9Ddt>uZeRH册I(!+\Ҩ EoQV,PUM 4AoUV(gXYR\a<+[*Is9Q%+vv)^¨ղōsgC|Y'kC_Cy,jf2Ym(_}~5~Ԭ@Rnnxxa :r zRh-ɌvtCPf`QAk a g?ŧi%\J ]*8 7ԢzD"քr"7/vygR(K͏oo{tX*NȋtYg><Q(CJ X-,"[k!_J?eȿZvGR4>ݑTnCQny+c0vUj#?J:Vy%N:$"-1vRѨ]ZThmhs>iµ!M3uZZ"(ib>t8Q{:kDY9+M/cF⊧KJv j1 J8;- 1 Jg!^ ~R ]ThҶ$T"FϻURIOOdLc(2Fcʶc*kdk X]QLF*^$laÇ%8l<jWE<1'Uq.(= IILV{vjW>iX!J<PG%:JhFzY@ өbV 6c,z? 6̼W3Wygߠ^9+¨gz8_)eh_VjzFe_Dז ,^KŪb$HjY_DFƕ#8?mEt;}y/vL2mv>ZII ,_9Ϣ1cG/g&`,8h?/r>+.RyzԦ:+be׼*l?D7H\_ωH4~뀈r1XcTc, !l0"yA^)))wAsFlޛ,Eyc֜0֦ܺ +#y22М^*J ΣBB`K2Dcæ"}-  ,gd X8)lW8D)y-^G1:ˑ(V,.S}L7~(]ykm4+*9tgt{㔄dEXH&_7{\L^pADW_Os{w_j*. Θ^127fBJF427Kd+npsCeN'RХQoļs!a՚.69er"M(*_A  +-3TF "`"ns`y# R bdK^ՁU#+Yu U!MZ=5d}KU\KWj+}Ҙ d 40IA([X#js85`T$hW`Q9\yDb+lwɵJnyq yʼnij)Y/Ûi}umtᇾ7o'}d+Kg)'w&RR'zV|?Lr>͛ߏCq{? tzw&r7>A f4JuЌ+?Q nmJG26ƾΛ,bwv]E${ v HlBr>u3ٺsdtVq(;ii34V9R 0g(z܀`!iNlY W~]O՜leىϰ"_dTGx<T=^G:?~`uI7Ԓ>Ųˊ)e}0V9ϩ ^\z4"ZCA.Y}$TMUIe@hpWAբ$gq鬮ENRHW }a*sj4y9ߒ&AUNI;I (/MYD,*A:`,ӜyIҨ3l9i/5D5id둲,jͿ`H+5l[H\> j0& 2ǹ)`%0b#F8v#s*M^!K ( "ZMqCQ*X ؍d#!`8/c7jD#锈RJS"*HU 4"!(jQEȰ"6 t_*O?@inxtjK@s:R<3<^Bm : Dǜo""\#<]c N3H\jՂѭd4SM^a~e5Yנ19,\תdjd Nu3,j)â eGKjȝj$Iz|3C<#Rz`RĭƙFGuWQ`ʝqǝb9^Rf0K*0$ DdO*_NwJgbjњ}8%_R3i݄f rFR ڼcX3ܩ?)f>1s\@pZh]-^US\gV-uKztO_iNɽZv{yM$r4 l_Sg薳y&xcnH,hσ+(bуglj˝f_岵U}hyˢ9R.H0# FFnq@._L?BZj;Y֤5cj/-B0$I@/Ѵͺq<&ɯW&3HWoگϸb (7ӨUK$ 1A~h-Q&\/^nnU TN~S46ӜΧPzsI$Ax_i2Y[/ҽ`pҦЛ )rxD0'iecݴ{dd s$*v>'E" XtDֹ !} , z?CA9܏>Ⱍ5qLmc7*tw?Z !Z!#qUy PKK+xA&7C`"&)6s]rѕ2K Oz[8dvj"kn4j^lA++)36*oI@{{?0W $q_Mv:)nj w߅Y&eZ0U]T2DKov$B|g70(u:)k9$Rتy pf2i/ TO`OGz`4 ӛQkm4<[$@^lF[{/^ThVN81SWv?=cYMS39kc)W`'{z dDM߃%efQ#q)Zt)u #G")L& /'Ty4Q{6ʵnH m$rѮ.9kï" añQslWIޮ ɽ Wv`k-~RiMu҇vO'bShMVkӬ"L""pn}3=XKU)Ѝdgl6] Ld)SS(q T? L)֥ 0S%M\>Mgh9B&|4՗AJpsn8Xa'h=5$z6E7LIO5L 0 FyLA&ZGxσf4g6[| j =,of[(C2f=a8uI;CE^u0 ]XطIE~>5.qtˤk\VunjHb%* pCfT4pYfq-}.A@icV9=Ӳ)؈a$hDP ۀ(QZ BQˆ+L"ǹ`%jaȳĚUƵq6fg j˾Yw3|n&ƞd0v[1DD(h'=JƞĽ[#sE(<@9PM>0#-u [aI$*`+Y!\^qԭƼ 0@`*M YR&xHD0 ֠O"#M'P]%ATtj\΂2Πu GF t[xk' 3WSPλ5T?3~snNK HGpVayt&_ nn6]}x70gŽ>لZ郞ՆKݟ>:wOs].E)hH+p9preoܧؤC'֖V/1XLP; Qg3 rfvo5hlt}پk 8Osm}~@SL^߹ |>Mp҂дŸ,bL{1 nɑPX& 9B=zos^-MY @˨Ƃ1J 8͙$ `wHZI\ll Hڊ)JY;.*z)j ؝/ڇH?k3XH͕; ~fƷ)l[\4 GTfrf]ʩۋF˗v&9+TIKfj/)՟JhTRd*ҚbhI X4_AB lw~Efz`+23tˌ{&>Q@Y|"t(ʚegYjd,B &gVMWq+J?oS&}84A_?n󏯮뿬[(fX/8XvUxw}yM+߾77WoW\GBaK:GŒ|~qs{-ƒЌYr8 ,N|eSdSOAB㣛vqgBc阥Ӝ:0uXw")vXM$ 4= ȁY"ORۻ bcs AC .L[>s );-wzK=3-cB'=9c7K>|Ӟ={XzԾY #kâ'3:Iz'NI p:gg䢓3 <G"5\֎=dgg➓ +N̮ 1EZ>f"h򩖚zr+*ߏߍ7\kU'lk*M::fҗϋCJNFY-QQ(4͔: 3$9o@5u!{{9ca±Z8;R^LҖpA+ l+0;XI/? }F ǠOw3X_9ƃi)5#MFhǂ|`sbh=q{3z8ӾEx=-Z {r?Bi [M>豏翿ī1i)@vz(7_۷گ2W >,-uK:/E?-V+~fJNeN%+Y޿|Qhb=N%8g*?vn*c'O7i8N_KNiOsV*A8Z#6/M|~6eVEY试%kd,y'wQAho| +a)&rޫ,yFʒn!QR׮ ƖkT\;ț/|1o[ i/o{WAn|r-h(d뉌H6 4Eb^=uƒP< rٖhۯ`hd20LsJpi%Hi=ov&Z@l/|Cӿmu֙Z6 ,:I7iX4NlV94X[\ZeܜHAQݫ^V\. ~ub̴޲E𡶄uȲqL]u/u}O8 Mߌ?Gsz,8ѳ@ U݆,Cb#P(f2B֥ZmJ@WC u>U={#i'вXvWr3뮆xqFf Izgj d2SGØc}籟NL~O8!,rNE~5z5ٺ-ȿN ӄA|s svojAJϪU)=Rz֕n?uHZ](. E2{2`k51*- ^Ѡ0iNJ9jQc-%f;[1vJ:R fVz^ &Clfo-`9GGpL,ٹ,5cgdak{:K*Er T&,*aQ[NջnNu\,RXwjDM,ڳǷ.);Sě޼!7oN'!00DJL0eD6ڏ=#dA ,&z:9s>Kz;0ޘii@(=Nh" )U ,Q B:V >*@( >#vZ{g?M{q`j95Ep{6WlJ9K]Кuy]ȥxʶ.J AwؘNf_V] Qd,菏Qnُ>3/]_d"(:pTنAMR7h=P`vY~&9r'=0 #Gҟlš5Ƃz>ցv dgdYճ*g]Yˡ1 kAՈNBp3ŽUI ʗ|FvK#._+Bia-ѝ\ES׎k&D͸Y;Axu`ˉ*`uI@N%FK6b ml9DYYkuuј|&ғX*Q:C{o'ĉV!)+0PE5z)%@o@ ΐFIr9z:kS!"hV)+YG!V @JAAڰc1|z$Ȃ%FqNծNjR4{Sfo=:VofRwcv5ok%EcFl/kqJmW[ Ѧ|ȻsQgPKH#'n$ӮܻdP.j-BBC"XoAm˥AvCU3=$B'PJD?JZT)^HTgn6({l pN(CKme)㖚Xtױk2 1|ͦBwTDo|EyCl74 Y*^GQFzO#dQaD=ʁS#TibU~1WeW:typ|~qs{~=\o>\%/.?\gV>i-{k@A}WWd՞_ܾz7=Fv%Sj&4k 'm?|CȯyLWcg#F!,֞S4 k_Ā)En{$sӛ4=w(#V]~BI%;Ks ;&s~5jcN%q-P7l1Lʹlxr%( fA[lORì (W ` OA| AA&!J,]*!.)ƣ P)y ;1ݦH/ZvOCZ(S^-,V^ O;;O|1b[ߐHlWA-2땒"C]:BCn= a`&GtmHӂﶿ>7Uo ^WMWl ŽRo<I#S Lc.wCڏ̛_?~389ٛ&u!~a "즩!oꨞ={<}{OfYTO>92r ډG>˩[Ixso:(h)3pPAS8Ju?ׯ-r5%bh1d|CJ`4`Mc8,RyA]/p44Y)OٻnkW4~霙fԸι4/9ٸI)&)[qI[/0IȦ |ZօHNX)we6sրFDH(#Ѯ`G6KY6)=d4%*9M$KWHٟX]l 2Y=C>IeXFi`21r,nODzNT{"x z=a#1u F^ m ^2dXcVH7ӆ4 a,YQ~.a{ta˅57NotnᄌY̙R:$~Ϯ1kDuHRf( IJj "s%ҊQB P2G\r*%jT^JŶK)9U$D)ыaP|7Vm@tw}'㔒0͌ǐ`o^~q  .»W!["D;,CcKJ)WRX6(VX; vY4eUː1B2 7v4) Kc%Js ǪU腼|ފYO/4"/T̉` r\"gz~BLӯ>]ޤWO(Ř32Gt G$(|PN"Fvf_r> "FmqTnblo5W֬|sP]/x kceRcb7{3Uhþ.s~6wc6u;K ~ysV5>,!w#͗D8<-O3`IZG&fan#{韏6c<lٔvÁV|fش], |DޔqL;^?nn£q욨Ӌf|E z<; wo~ifo#v=([o2v?a& 'qa~]Xh69[TZMdH"&(KW!"qZ*?i -'$>m0e,V4xנԲ0 .ÍrX%=f@?mCo%3j{%8…JgA<9_!-Z!aܾu'å9|t=]F KZ0C2H8T) fRJ$< J: I AJH+k2fY!`eC1[4F $AՊ2j2pJqr~hU?O?ҜI+.PR[䂠6"YSy6x!m 8i{xj 'Vx@lTw9FUŊ:KIbKnQ+Gmܥ&vgZ۴R)8UlOȲRpEx8ŝj;l?kb\ y>圆]{sDb\F0c 6׍!OOXAs !|RQ|<2`gj!|h ltP(p*h)՜@$),,eBYbHXK ƍGD+1*lfB@WI Vt]T?7n;6ldUXd.N!p#UTh6˖BF6 j1BiWJYL8mIBV";W韇{'hA\+CDP:BԉQZ$ XߏY#ū eiG64AR@6|3Ŕӆ ^ub1 Nl h'H!'tؤKDZjXфIh0\K 9ORFcLWRhx Y(>MR˖R)XCKd={낖c,H_ ^E˾FVSNl{"猳zE|;h^tH>z܁*v!I`TWzrMէS{סvqw8X>_Q+B)E݊t/c!*:Nr5: ' 8wr7ڹMβ쁄|$?7#]"j@~D?5THLQ)D&`Hh C oP!s:z%吡`+_TS0ߺlVy'0%|!O!yJt'3=S67>?&$l\MZYkccq|I,hdfac99 [Bs\QWzÏ2eݛ[vHPC˶ppF8pΫV2g (T60kJ kpPPXGq<]1Tqst+1oFC8 j|FpV+Khl DKr=nF30d)&;NS?jÑ#>bM $:.UTŀ-HvwؑK\ E;vQ#H",^[ &sWvR7rd2@4JrKhpH %LmHqR~t l t-I8fΧMγO>Q01Kg9GŚ3wqw/޽[b1złot]U݌5_,'p}qq?!K[ė *Cg."U6 zLa&j|_KkOpؤ $tB2۽%W*םpY}YD(]ZsCL >|IX2ɔs8o P+АX~)8Brm?vZ#Y#Yd)d,d]&^7 W S13#=W-SfpX' 0KG^t ($&{iLrei kћ2wYOUgHQ" wddGn0T`09zE8/=Fmeny%1\@R8!|q I8YS ݍ<ܯEZ uZ<'YK1fȠ~БDtB1*#'A~-B$qVìH|x 3[rNDxB+ HGujf0XhQر+*JY0q;Ӭ~EDtnw7DṊc ( HwJ]e=קu?LnT'?:5h @`@[f0Z!ce1 )aQPQn9H_J7>99"%#"-ן|wΝSZl5*n`Olg_v f#9Ћ 69_8Ө\}9v+].E=@.>!KE5|y^Ho/Тs8pm| YՊ< &'TY?OXLK@ 2@{k.)Kc<S~aORs_nL{g:DE R1j&b`Ud7a Gu èP~pEh YYGIhtY8ctƌLDzkȂa# ?yuPP̊_$G ߰,/#A-ԇq!Z t)[/e0n:n^8,>:[9_?\Ŕ¤lBq kSXT-@:'O"rƳggTN_/A!<ŃgDΎbӪenM%5F4 "GJDE|غ[hƇuG`9f y=$я\a7MCg%8 3Hh~)wEoB#V9xkSW4*p)0)9mH fZ :”H5@9g,j1; N±U"łUh ьQ )s$Ǧ\7z@@6Dơ x˱Ii;h[ La98 W 1F6i-6ϡnC}4qkũpO4ד,^zO|޿[M`٨WFٻ֦6fWT|IR3!$Tl[ۂb! Im$Xq}ycjwg{zyzgma( 5fw$NoN.%z\Zm0 O[+24@НɡLk#ˑمbT)j:,W~践qudC]0~+C5 ]ݸQ<\3|xki6r9a c1kGʛֵhAr9 PˉXB3ۦD wf\X\NN8+׼ۣυ+Y!fJ14JlҮn]L$]J~. .d))S`WUhC8uhk9[ujQ~#`U|5 -~67_J"jeKfeU J%)YNw0J>Ӵr]c}pJ&",$i*X9MQQrv ӋOhUi:eoԉ,c=શK`N&D- .'quA1nϓ=<-JTx㒖r?Hoa~-1Su9dZ%ZEJ/.Tm)ՔVIHgO'UtY"/|O̱f[o="J.Dh6T\lE~ryL۫U׫L I J}@*1@ZKI8d2UrO'u7*IfşiOSvӼ|ͳt0c(a$nN0t"6&_l?KCSqJ*UnuIX+͑(5xp\OnZN";_ qV@2 #SgFS)N"aƘ ޕZ6&/|SHG ;\Y,ՎVڡpH+P*4!Rrq/Yբ9\m]6I/r[b.9I^ ڶH\J!>k[Óĵ<82aB9=,}>=ELׂWX@E`\x_g>'$y  j) ]*Vs"YՎW/OrjXa^zcJT$ (q(X;'不rHV J$S^E\5XFc&L:xaܸ5y_~{g^J?GVWL꽧fP!v{/[BLeͩs>L[7$y.e)er\RʈeXT߽u+⸔Bk\LJ&|r"|_`N]32fΕ:jB_)W#Je˧Ql\N% h¬4&@FK&in'V1Q"v5fxG\RJޘʦ*KnR X߱3ݵc ӖH}KA)N%D)8)ags޹)^_% ή]d%q.i%k `vҔ;g$0()a7oULQ @ 7ku#XBǸ+CLq.GVj0Z[J%Xg*`w(Ѩ"ף$&. PNU ^#LVu'9$KfA^^r8 'i~T N3n`)C%'QƜ׫(nR C鬢䂽U ;^x%%EbOh e'W.dSKI1sjq`hIS^?vPAU)G)qq98R!q}x4O1;M\O|K oN|vp:É_@ t\c K 0HbcdNZJzm}K/q7iuBIM34VQ 3ĝR>H1f 7Z* @Ek /pzRF.2৻;+|<8%*PxHSKf^E !H{4AZÕ6p3'k9qaW{%]`=vw M&H*`X|u  zLyYk iI J{pUd(ԻJRw `NMpWDÁ:"e$0Mg`/(3(Я@HÀ[eTjRR'lr2%rm tiQ(a5ģ璋2E&8hT,#Z2ĸHX[Y2.0)V~ oM5d IjX a߸d`?6~awJ{#[ ?/jF6d~F^->&v>2+R hz;C htouR&dpxj &G-f)q 6,6SHuCf?M}k&L#me&Fц x 򍐵rl1<F/ZFn|a*Y8GbE RD|V#б|%J gܝj! bIP,V$][bՋ[P^RڂmHYPa; > [P2C%_ZH^hn~TH2Lp3[Т E_| 4~ CQeȱHrh{摂Ve&^Dž&RߘHNa@$oXE+ڵzu^ ַON3%ұk?,3wplM El{ Qsw S v0c,:}{R_??4ovxJxrS&x}t`6YM-Yw]m7fn6׿VZ4;[{;{7mnZӴBs~jt vK+^Hv$0|gi7zݮ7 :h^\+m3×ɱQ܁LWb:dy?mpvDl5{9>_`o?lkR{ m?|wU8hN g~V^apu^5^YqyIj$[^ h:il:1o'ڨ`Aקs޵?:% ݭ=- Voh7m:fFv2~5rdxBe+?׭NKs? {5 ~`[+lvneh}`/8pDo` .!;OiMzJڣOaZz=Һ.iwFrNwQ DSYZy|u3n [0^?]nx&Hkۏ nu (Qlmu>2Hћ7:v2L)9^Fa^\55p=FS'ۯY/;ze >}lܽ vݓ=ma}[^3{pmf_c.G}&rl+eƮV'sNpxn83rt52FIzEb^4reN&}25-k{m&flo4SLR5%'m~A@$: >i9a Pʍ$-f)nm )n X2-)6B8:M~P$m}k$q0qҋ6%`=?5RΘl؛:5B@itc:*D{փpOk(g\T;b"Wٯ2o?\ٵpˑ5_K緌51f.o4=y)g  f@u~M vφWGwC02-e8=gp}ϯ}`ӷ= ػ~]Tm^f=uf;rON0E5#]NB` )g.pH``]p)+!T\ /~EfD:,.-r68ŸR6pv CHgoa,ELfJ Pi8GE,p™R5CKbOjP3! fcUw&H%{0@3&'k/v%4Iqo i+̢"Zv ILi BD>F9xVD56-IcOqoC5~Qw-S]Vװ5^^"QROq 3$Vhy]%g)QHRZ4\k}PF53eH`nU RaX*ׂBYLaUȥ;-S1e p'NVW;kVXA/΄Ԭf|~ V Aʝ|d1CI{Nb J}@*1\ㄥR$ Drw;R'"$v^i%)í$b@ǤV adl>p^sv=tύe4όet,2f : 9F fn B3l P_ؙkym:<6qhȺ$JlVX#!6^cY$BcJ;b$AiUae` ឍ/(UT5a^\T'Tnd 8H hAjK`24ּᥐ$*!޻|IK7|PCMwmcur]ZrsMOI>!nfZ!WL"ގȍuB$x˚'Xۮ!ݽ+X1lBļEDC:)!Ԛmo`V!&T2ر,J\#L[+T~Y##s&EEo$ pPH3 ʘscvb*:%oj[ؿV?1C baN;cY1U7*JRڎB5Yu\1D:QktZU$_xa/CjTbd Dظ±VLrh`or$`CRY|@ UǶ!FiO23ʙUdG#e@X^z>#;:n{>m"?VZEV1i}IkAE$<(3S 9sr_`9-foA! X2[p/Ʒ_V H, '{' f".ѭ7բ{i.Xmxb.>߂7V߂0 dy9CkloAnV՝K}u@Н)@mm Y[7ӣB[>}nm 7U lEc__Gߛ[l[~얖X^O/- [~\BμeX/垙X)Te~vJ\t=A̻`&Wn H<>*WK|g@(Tv³+T#gْVdEa8}},T>:WݱP rπg}jJߛ{MOP4]KjCi#o(TKz-+T~cgwf\MmMRl9' EZm^$#I.!fH]uMwFn?A/N^ ;m5RKct; RQUJЕ%r"ƊR8z5k1IUW]7 B`enȵyc'Hۍ-S4V CZ<]MYGSzeX}eD0ix!8نXy֣0%!(Nm2u/Kye^}vl}7lm-} 5 +k%xXCv2q8Pp`7ԽB,fv C\PVR# 6Ks .P\FQ \-@V\6fWH66u]ngٵLtY&980i!oG|UjTH9Bs58bXœȮʲSt"m/ZKmiMΒG6fL)Q<0 :o1lXY:S %䷙^PL:!Q 6C-~LY>Ԩ*Q4ܜ;QϿz[(R*>6ncO6|)o3oA]>_S;ɀo/yo翁]޼8ݟ?Z8=߾h>n;̵֑|@_=8K3rf4NP6*7:=y }qG5xg?KZ)9=٭g?JSuǑZƨ}"'\*X:b[RD^jK r)66_j ~)=hRAH|\k ڞ< S.+Խ֮;Pkڼq^J܄àrv4,^3p'*\NSmJn &Pɭ9 0D^P-+ QPHPA3QIS T6m0\xdLwuٰ10Zgvā@ 1҇K4K îxlP,qy"~VWƄE:S4gD+Vy'*`6whx]Q0 BtlE떼wҵ[>LfU0RC=Q^9 ^0\s4 J <)_%<}P:b7-HzY.Pe~T;j[ eÐ5lP HwW .Oivbb =S0(-{+}ҶCGi=o{Lf+# +N>T@XWJd LrH.#~;$:8{dl )X ԫ$mN[w~z`i ^q BB,̤W|(pBXN(YzdJlAx C_i|U‹k <$mZb2XSN&BN'N<CI `j@աG8p ф Xf\:)Gg 7@]Y% JMbm^`^wꡈmH;9T$_v\Z ݙxLo_No$.Ҫ0=QܤH%"qz* CFNz>wD \Zۊ=;É{dv=',p2dH4imC&qٺH xk7.;P^ˀӿ<δDCkݻM05.97 ΄WOGʭM10e ? QA&0`/,;I#u K? :=Pr;8Cùic;dFSm9o8GO'mצݙl6]褗z'ۏ5~9.ɂ^+wywdD:9oaˏsd/@xPb `KI:5T~cM+~wˇ3/W\Zoo_8zy7?w~<_O.󛓛E͗ORra !.FqOE>uN_>ig^^)Zʾdϳ^ 8xe(5+i2ϖl}f^N>{9!L>0_p?>d fjÔ]pd~3p[?M޽Ϲr4OyL3%2_`>y/Jz䓌i.t:;K0mtNE|z:v 侅}2?6`VMu8d(O:p)n Shr7ۇQen^qRۋTOUqZ;^..̸l5Z~ƱŽ'nXfq϶ԎbTsgض;qD9s>Mf(t^Yއsu,7!NX<1ޕ!\4{zԺ& ,9A|ܿyZdhO7^ˁ7q=+0(4L:hAш[ޥv<(a?iij'?5̲7q i,+W5dW"ˋF;hp00L]Xz}6s\tgmpD?ߎO}?i}?Mn_Ë?_=;}OӼ1O =s~g/Oϐo=_g'NO>p'O~34{Ͻ=;<:9rķ:nn華=/z;۹i8B~_jp|];Zgٗkަ;ca1_;W@,"<|kPrŽ9j:5]TN wG52)2?u0ռm7~?~fWM14AT=KBQ<=5(w)' |~RBƣ]PɃӳzu?eSOj_;TKo^abToGI_]{xI|W?k4}^Dz BMtfwCIZxutpϓrag$vMS~SAU*oJ ?Ԟ[xP>8R |_=q*BvgeL|lѺQfNɋ&TףɣC1Ɵ/~yz>[|$λ-ߧǯ^k C)_i-WXWw`He91@3&g2&VhQ{ y0s%[AJ¦'~^#!G0 ZekEG947{I#] ~)_8TfD_MUUFG#o? j^t^<}g=s!1R+ P& WVи'AsNX ƪ&g&llX;c9'Ӊ-arv,gr~@s7qƎ|w3 R4og명<k 9ƐYd$KO$W øD4K]{3Ș1&˅QVJ2;k$x/%:rዓã!ř`I܇ƣEyٳ#x-xޙ䙠$"c8Jmo.8eJ̎0;r?)/UR:,7 nmnЕ ai1}4ꗤ6267Xev\c)b 2ɔCX !Do*ڞU+:%S+x[*"Vc7 okOFfw'L N1YZRXK)VރvGYaoB1qeLqG\N&`E;@8ǭS쉠K8bª ńJ,X Fa_ڨhz"@̒Qnբ]`tWr*s_[Y{UM)) *shpTlæMs?{7{W~`zFYNm4?pӋ㑩fmw{%x1phZPk"j}9vAH2A˘uu$%Xq`oaɃ=f&mC5kbM#m M(h*I| D夅XF%zu;'@`TwCaEx@ rLJ44\TaV-< PL`xœɽ/b`j`9kf3ffL:}nԚBq`"NS{icy]ipkGYrÁ@he]ܪ(>Q4{FL<1Bc-Hb[Aj)rQuDB{V7)SI縘d7;u$ $:G10Ci52X! @:hcPZ X,C9T}L;ӹCr1uf5`K1D 5 LL. z]5zȶv\ܢ⺗X8R=R4 ? !ΰ5!)9`Ǜ7o48GJF*N$4 AMn$7݄T#=bF_G߳盺ؗS0N፯j܋ FT0wd!̵eLdĽNȂUgU諯`8q|Z2^ShkGX<ϝ"g*kyeȭn=j WbV;N}_,nXE^_V.F"- 3.cT\ĸ$wV1Ը:Ybp6ZG 9nC%ۇ2TTj&j&VɲrE~jFXp[p6xRg1=|G*aN.2vHJ E9X̴௹.q쫋%܊II"a{]d dg_Lэec07|{-~'\e,=-إT,1\D!]^|e=`սkW] 6)ĀY^@lurw|x/F! G(`{a67'3Do y- &wߛ %*n=\70*~Rs\yԄ DrYj-i.*IO3JDC a*DoUsYQP\x@'JD-=Wl#+3p /<-w[!_^L(,ގ[W;U^mݵG8q[{ht&Yz}sߡ6~5?_)| FT~fW>]c'_o=9'd}#ؓS`+3YN?=8V@Da6`~\}Ly ]gcog)(zȶRF|2 5#&F {SK==|Xd+~qJ~2-7K~}Z ´pw8<_(=iP-9'x70^ȕVh ~Y} hm>x.T}su=3֣['o):9zBQ7Mg4بpX)Wcbmmj1KTo,z_rA.wEAw}yi"pZS)Mfzom&Bj*W*/"L _,X%&)|C>ARGD H|ͤc)*zݏ+l~E b婧뵛179l/n>]_?,z]2Pn4I1>lχ2"_|Nn>yIGktRڅ 7v|An7do{-p2S%HIXM@*PtWQьuR#TQRVvc~5Sԣߪw9U;H DS? &{6e_< `5k֟$jX,'r Z}eRû.+=~zW|Q.FGWX`#+B8aݢA][ݢHZ9,)tg]{ny^#۸(h{p7c߯vJ@d6ԤBL?]v0[hKRk s'JXNx?qEJ$S5yMdFWTkگjSK;7ɥtw׋~^n\2NԵ[A=y[~s~|G-hzvmY-Dfcԧu=8,/㭴.50 RodL+'i]jU:[Y) ,եea[:e+?Z~xi@_ӺłfwիyLrk';BRJ\<^r~pO6n_wss8Vi!\S>eC$IV JAlR,qjꅝPѭcJ!jG| .p*3;D>YAJu,VD?ط%M37%422Q`(i94*KB>2=ET]oS3ÓA0ƋK=X An}PQz~ghXj3;8dAf̞*;E1k>LH2 (0%=bׄDcOpC>t>UِkU2>.>[ arBk*4J 2rmJm54I@EҊ2H\.Rf+$fWE$:T)=Fٜ(0L|W%d vrMܐjmY"Br?T_\"@ V09fS^j1F!&)SgZ@-Jpzb 㟋[ TZ9#\4S:yfR\No9wܰi&& D'hh@V鐶EzJRj{B|ie&G))s{ lf#PNs\{+NYX Lh |>r12O#: n}yQkƧOMv&)Pz ܀QbxC[޸!6$v dG# (_C!w8DYQLȐs+7udۃ v)1tȭCtQJU9KAcP.a'mTX jjQM]BkjfX(3'(l0O=<OP$ZR# iOv1: BT9OРs Hr'(~M pn4<'(?%x,A"(yI]IH:BU͒8Y5m޼ׄB!]Mus?/H5B\&^9:HM3]\jᦾ#9K9}څ?E ªmwnW Xlkwo+r|⇫,&sήe%颩:-jY-ReZ7/ϯM# NfqpqEPVx?6[ۻgB+|:?Gs[̝ud!D[ٔ$_Mq ޭBL7bۈÃyW$[MMU뼧q>nTwki6.a*`ޭ}1FwB^6)F!:CyV>d?d*OӺ|I/)C`~ RjVOӺn VzVV*&b>KMiEJJE`Y( |IX,K*X g+=.+eB?WZR֥B}:JY2HxG;nyJ?קyžօ.UT>e-CIfX#4p@Ha ${ cϒѠQI7!yv'Ǡh&T?o)Ye&Yrejlm>]| ^& B#_\1A@K aAm+u=bX4 0-x#pn) PPzb >;˕" x o6gI~!#X®9NK88 ԡo\38cYیPG{aZ_-I+满v*Cް,偗X x¸Hb_x[}žքDV./vPI6x^QCx Vǂ-A(XeF6 ,YU"Aa\xخ:(JTIWW3t<.2qj&$Ki3Ќڥms)]C.5VRY!9\lփyi%4BDcHc]:{Ij ٫bL50f058> Z82H9c3żVGa|,>MzMt.W +>&=\^|^eQraQpKʔgj{ȑ_K\4&d(l]Y8eHC6!{FrOϋf4ΖŪ"YTYU(J[RjDJYMFqY8H5TQ^p dmC,uWs[ʼT{V4d*m&X*X V̩U%=5RiIFt0lm,hd2m^e(3ˀseRrk(0?1T ٘6YjÍ15XT^GHZOY. EDanWEKj=dqI\ű x'[r^ y*WMhɉINNo~XJcHVזt?0Nqxl{𥆘)[basz<8{NIړԅ缳CUM=Ƥ:ܳō1&p.:¥aJ [ۯyN6?wخ4$GxT`81=Ax(#FBZΫUBQhij '8ʑͰ>rfj&HT7%؇TۢޠG+<@3F F tw,Ga1zby9D8@4[U`?j4f` qSdiciq*h~}hEP3i^CCw:+M]jkI|YjݤSF BtetIIZc^!n-KEm\g-57\ T"~|XTw.' ¡-~ksw:}X{^p.oKW},Ӹt!U~ywݾ ϏHt_^هG.$zXƫz6g>~m6ں)}֏6՗nݫȡ}7nrsB8D0eW}=zT@'1mٴ &[yMGލ :߈nz$6wK/n} C)q͂hUlaGϧBp J^v^OKړIX>eq) |:oXMSR6V(" B^YԫԠ e#Je/Wz(xJ5 9KYr:z'5/_Pzn(* \m=a(* М[#JP 轒@)Wm5H.(=G CSj+0Rs4JRa(UY,PzJU[jAtR{R}s@ =mٺܻP{R}ۆY4lcsD`a(lsg}ʎ>F:h$JALY!F 8YRr[X. *PټD Bu[Eu}P4%zO5MtboxiJ?ߠrŗW'hoÆfZO{{\}\xv_"'QduuD8)mjoU~W!Aɚ74wӃzxOYz|x |Kdo,n 4mAT`+t|4#9\?e-jtT`G)M.>[# 2@ CҀ]|]p#p?ow&zM~<89rn[8즶onz{)4j!3 CxuD!Z][*sϭsFÍ䩦4 ZTrY dK*(d(%s΁Jn8jmv Y|v.1vXd Kk-5!- GUǿqoٺAa>t -alމ"dX }A0KujhSo2S%~&-3b_’0› _?;lzmno*٢t͙凰|q?*wFoѿr|B'O<"?5}AqswmRTCPTe嬁E&3*V,}i??f_/6m7!C׋-t];R?vM:0ʸB5w=3nBʙ݅2Bs!jE 02 +eԑԃx 1l޽aRKɪ=EwE: tFdh|Λ'[[FyۧZ8E9dUxxWW_:PLҩ^y&x,XX_TGY\Y3Q r\ Y35ꐖRIH&ˍS2r{ǐ TYJhkpڷ$a(_8%@YYLfFjQ$+2ᴉYI٥ztOO^qCC}zQ>nvI?|KF 0~ONkFn򆾻_Wo{| ʷ=oIr/~407}i.n[tut "~,IIhߧn|q:>0hf_>F%m4?I^ ݄oy 0 -Ft75;85.o'KX€zS B =!j8φiTvpc]-Z!~Ύ9\ B@gl$FՆ3 -'z /pe^rSՅcYrs5oUZ([3:k s=G{bg=lYzXoz g|hk;V<,)>kWrvڸ]Šl.ERI؂_XR^M6a0Ö ~ 71s;_2˓0i@3rdf'Gs/lmRNCe½V\3Y.l%Ī 8G mRFL"(KB sU\RN{ EBU©eY4վM͒9j.&\R}X -f󞋥 B[Ό[բNP`:[%:r)Z$8-Ps X"7@ :‹a,E)F6خxw˂Iu=tHQJu#@8pkbcH[{ "[e׌;\8J1vI lU/nE!0Zr8\!t(tt%4LE'"b:Lpvfėr`|N}/$R4ܒvU0?5ϣK3EB%zx^֒{]tw.6Onb"ʢ3K$vϊp.W1$SUB}-FX"x6fɥj2~=Rԑ6)Rg{^1-]՟x/]jGAWRq\ [jFwqd'nNH$oN8xjH6= $!oIJS3:L`R0}#λM{(hD|!)8r1[B6.B5[ [yŔL̳ ')W{RSم8aD K` 4UEȩV!o8ت/l)*qSԫn/S ,S[hN@)WmluP*YP[$fZf[u2e{qZ՞M(8;ebzU>5Qp<&iE$y^B b ]2)nXΧ0JC֨I&z4R:r|=>"Ն/Z7F(e6Db<mǀ87(/,VSQJ*J-$VyPRaY0B _9)Ql*Hw/UZ%k6[[ $eB1f(ֺTĉWdUVhJe9=5@5ʥ0P{62"SDJVFRϢdUt xp̵'fc ̱󚍩luK?YFV0_4ښbEQR(t32WED`h I2] 0= Ki7r[;)ܖuD$D{4nB\!ܖ}j}JrZ{) Aq[n-6Wl)QB<6=Rb A+=Hr&6vZ՞Ԋ.ic睐;ِ5t5͏0-L:n+uHFHXrRPoW,%*iXO6?wخDFyXN},^`Ѹ\2[|~ԕiaǢ)j 꼡n'8sGbKptYbF&_ H.X"pЃqkn{(u{Vwۤvc֣Ub+& UkUS\85lZ g-@d?XWOnb bρPά%XdzA@Qj7l)jp2iAb 2m|r0% 󎼝>ܷziCEpŋ;ۑ:} OZJzK6u@#J;ٽ&~za5F^@0փȫ,|ڨTΙ},JmU|J^a`p߈N6>6蕞 I>eG[݄-U!6aExm.SJ>܎nu:߈n"%9^tK@$r.Ö_\a'RV[Jcm*ٔ_# t`,+RU#Y9 /U楍 M<.&eWj:a'/=./@"]\moٻې^ɵԠ9yQ{0/ V{ GZp|HehaGtG*K5(^g/)h^:gW͍zxYy޲Im]RZ1}oJ]|Xt|kG#JXVal6aAQ{^d,lYE!-  * ;YPXVe`9\ru8toIf Rɂ* ?;K97C~a-X-?,sVew}3u2( ԗ]%*:[*p WX{uu/f'/=j/$ Rf^楍>S;K)0O!SUيYx)0/ð?yqy6K{u/RK 9q{)`Bi4R0/mmKK(z{FYxiz-\d\v^J(5,?@90 Ra~d!HXeȕ!Gʢq0 \O*4ȼ*rg sKAUmtVWl_?onͼ.eGX=() {:} 8 1M맇3Q%=CNdN󡓪EVB,0/slV_ܧֱِwCyp G/ȵwb" 5X*^Awa.6C P+!&\s[T[RGíu0H{a逤3< y@J)We̹RV5hb|&_{5T/{׫+8}qǫ-]=]?! ?#&K>s0eE:+䜛Jkk*+eZyfl $eYV#X-kSiɄT7co ;ꐗl$>z /t 0˅rB\jM,s9V*AƇpT*v uAJix%6L7KBYgM^lԃ>rA~^Ђ,wGb1 c-ZIgT/=z)DKRv 8ozSvvbmltb?wgͿ5Y҃7{|IOy>"W]?~e6xaڰ^FKPr jn^LFK=8ܯ^XF5K,},OPbq2(&@ӣ[q,JZ3x?^V 0~D6cHr[BlPXQӠ*DrFIX4t}q*0@axHz:z8RZӤdcDS{fMFN!P<ֳW-},J%%7?6gOgzmۧC+wE:i}[wqt]hW7>J_/dzw~|Qu͏|J8񡶩 `7Y =>לƐm瑼#x@]42jm .S(`#vtv薪IFLt.mh-(Z`tC>&T_pֈVt#ں2wKAmY"r-Ch6|pTpj5b2–_l ~n)I+:\S[sRG-GWS̶V/.K{W Ѻ8o>(:9|ZKO󯏏w>IO>w־Y^eg޸/zzC㿋jߖ|NP2S̋zW=7&uO|Z?U94@[`Q'v(%lSO?Ó >sMo5?9|G+X89L*jwk7PBwoQ [lmL~nնvՋ R2vt.[ wz#\_kJp21L;αq ;K0ӾAEjY.I3E*D}9a4$8V(#L+(K?^PV/R R 9kKE)..tUʖJTd`["ؓr!]5{lyxz8<򆤗]Ihغʑ{%M`ވf!,ݻIFiF4J&0I=CI̤d&8Sl"8u>k sU`AUU-|Em,;[uպ'7C{7d}( 콠~QbCMu\A&6WPԖh];9,?­̧\QsШd['þ0'YO%Ťm(yV M= Hi5(2*w4tf4.'}n @?JYT"UġFPZKVo) Ox*C+ ҹ,+ (m !)tRۂ5_+ YJ3dWa:Q,wC09"2KJR+,yJ*ˌPrv[:T:TKf˒7E! T -}b%E& Kf `΍! Hσ&]ulbAWJ?.}RDJ6"z >'NJ ])bN y74cG5 0[ 2`_=8/y_}r鋳˄So_u5͙|̿n>M2hA#>? e&~`OQ>~(}@ƠQ8~sr`U"L(ꃿ{cЂ< %>jiRk *R7kfZPvD|}a8hkx*:1\ ӣn1`YQImGݑ@j1X`<?D8'hA\`m,B 3QsߖQ c2z#N(̱+pHJ<5\02lJ%' 7yJ)qSO{)z*hxaYIf.V :J~#y:a<1O@]4ŧFTN7bۄyW}XZF!<䃻hO}.ڕ&n:qoD gަzZCW6]S2ЧRZc|%֚_Շa *<ӌ(8fdw'4cP2J.IKh֏_}1כkfzٕt+2 L|fA{!Fn2khILҿ &ajWLiAQ.8uBf|pҿ $ ~qu}ጧC?W8t&^ Z䚲Rg2nȖ[g|C> зMX״4 'QX# B- s ~dkX=p4YFt %\_wR\ul0BƄXA arl,c W-(ٚN$ Ξs&U@Wh)*JQ2 (̬ }3\Y&F ":5BTUTeB}*xy2֕jU2%yʠr&GC%RXHZmi>t2}h17@<\])H_$cFȖPYE "-7"ZrG("BTwtl}1#kD^UJcY)#i3̈T.Ԅ ^dj2EF^8Hz %4Q2R->VT&׍Ob(3}g_2J+$(g 0*ʎ J-_b0-#hw;o/#NS LJ5j+#ٻ6r%W[ego]@ dđYIvDj*H5}хFeR)Ѩ*iT@O^''MIuѭ j]S;G*!Siq>nx[_@;D}V [Ǝn TF!CtSSfBH14^ݪ& ']z󗹿yW"Ͽs^f'G~wE=Xo\58^|[=mf4FɲQoܯ?Mmowa o?>PcJpXGumc8!$?$0D.ܐzuhRl{zYH=e۞Pm6IZQ:r A9u($Ԏu'5JNJ7I}*5'5Jա\JzۣtR iN%1TUևRJ-Pz(t T?JCFb鑣VNp.WVk|ǧ7 {*D3hAXz_۾g*:hÅV{oE^ǝ{'岀00-d>͂LbvH9:g"y#bcK|綨r3Aվvͮ,UMiDN>Шw:?vp?sܪ79JJ۫x (.ZFj|u'DŹ*u@{.U T *V+G 𶑾1;)p+EC6@MT#J B{m2>w~ qwU{hok QIDDpB2xʖ'!2d=/Dtd)njptU_.+/oD*kޥޜ%iR-.mR/qUZܹVĔb^4W\wԖi;TR{"6Ucd`T#Lh1I &F) >N&Gz[s5&J,Yݏ1)g4E+Q0ch!9gLN؀A E0RՇ"TTmyNRKAxyWҌӻoyErT>"5TJ E7z u3 5OX"Au`ٴRjrɉ2kxëm J>n|UjIH"i؝Pz(J -3e{Ҝ93u7͢ddo󪛉:ߵ/ mGs%ָEq:jNNcָ]K%C[.MgmK梨Mzh ţP2 g}sw/bksزsU(x&_~uɜ5[1e^槌Oo#],Iڼc+<>5ZJ5L͘VлcGl~(N$52 9F樥b^õN4<|=Ϙ1q돷'<;IJcviXzӀ kvDcyUukӨzdh6Nب,rD=+*[]Zմ~?χUhBf;/)X%K*s~d!,R&-a9-XN ,mX?!/X(c<ܠJ&oOwIXPB8]?Gץ 9[e/zs뒛LsMpK˷e|yua|J4B㐽+,uM&kǩPuFj"uVF! yc*ddHQ94M!$ڋ-^FtGlDPʣaQgH,bP!蔰Y([$F`G%ɪ1:,11Fr^֬2K-M"k9vȺ '>*SGEv/9=X pN]n?3J^u|oyYR?ڞ#Gu(b"e!R~;fX oC*u/Won.],}j4y~ }qv( )ٻk|o U im=s DLک5믐JRKXvd6%¨UqʕT"6OAŚ[areѸL{}N)2!^͌f K- ן/dqZ\'?0'V4@eNhJ Z:-q1k 1 -KKJ<2KeX˵DJ677Q/3ǩ .%R}s|C@2hț + 'Xe2K?&@}jfntrcEjd"h1g9plXw/6{b{N*rsj i"jwWoWKkũNzA[.Y~, UΘx7Ɗ1s99)E7fPe(½,DGsaR^b~5]1-!i6 )4U92FqI76 6Q9GӞ{iLжMRɔ2`"& ziѴY\6|cB09 O%(K+fTLD3'[iQL3lv ihj?Qwrk 3400PS Q@=-drB@7YrNYѥԧ[՗#bl u>mz0F ]dub>цtW LO6]ŅrB  Pp륝h_re3Jc#elBJC/|?<}L  Lif[ط7Sţ <F0U=PfN)1撃׉h3~fRei)0fX9vI "wf0j,rlC%쇂Y j1Sm}X|ht=:Q8w]~SaE\'7У o9z>LAѼ0+(a4Lu?6?: U'2T := -sEsqGk8eKas;&=RP筑,rO$i< }-xοyac!džGwf Z5Ĭ%fM3`4h9 ڀrcrþ]OxBEe!܋@!wF ɮTyO8"\yԢΫyUY+Di>QIU1'ޟH򀛄Q2MMjͮ0 MQ3Sl"(uZ4/ʊ1V{N3LcIP㙦J5Lk~?MU*W4ͩLFSv% ٦I"CƸ#'®W֜&lGGU_WWZCE΢=ԏi3M?0flMQP1)wb1hӣҲ )0[+)4_t1 Kdmjhu:Ăʖm03뭺¸MF9;M;z T&q+"E/|"-{zܞS 䒱[V$9izqJ%ݕd[)Dj3uʓ] PQ$tc$!о6; ?;Ou~qWb\ښ zֳ{,]ʻ= %;[ic ݢLG[.Qi YDMAD΃J8X~34,Ҫc2+%ukf &#:DdὊ3QRަ)dgoo34 ')הiƹ~:-X,Ո֊`g$;g 'q|37 4Ib歮>bB͐}}ޏ/gg93߳1S7"@7"fN#j jnSan6:Ĉ ]Ĭ+ȍL{Lg8Bъh7k-I`o$8/70?Qsb)[H6s5gW⦅)7 ~?^rwdJn]Ն5alcN*2 "b0#6icaS~`z^G=>mr4FO h29ԩp$o,8,5YA>Q 9ؖrsӖrMTN]YVas]Z2L>MUSbY}7;2AAzx9M=r׏wߏ`\{6 }Ԃij~[u'^t5(|/Fwh[j?}rZu?J5:re~բwkGcfOo{wI/>c>uyy&%Lƪ", 4Qlm}>+&^t9˘vmb7[:{9´B<ߡ% y*H@Jк)Ga Ku>u; ?8i8׺!/\EGq>n@>T9u!ͺ5?FiА;#*'`:}~nM&uzslҞQz<:;x]\K{G5Sh4H|<--gg/ׯI֑tݼ1>wFmS>f_ST$91GFnu< :iSR򴔱>SZ9M0tɷ4\ *󥑺@ߓc7[Wp1pC __ i6ЭElkaR dcOzRL>V^-~!0BRhB :Q~1iK3ih  |lChp5xKMUg}B%T.e4{6* SBѲznd1 gώ,=C@;U*g݁)q Tj@ mQBK5'G'-es*ŕPLt6Ys !(MtIJe߬1vtJ:^Ҷ9)}4jIIyqkD02b$]AsDKNII[b>pO!AEuYu.l `S(XxY/YFi:< A38=-om-.]srAB" РO&`'!LHRVwxU<&ri:CaFFP CJqqkT/ H [GaL>fn]ω%%pA8I%`ʔgG<)D90aJy.X٨Rֈ R\eqhtZ8Ei`1$$ *JLT^ Xé9r T t&s*p_Tιx)Ha}`_rkzϚ*uVo(R >*2X`h!=KbR>A5CMsPΤp,ILk5)\rS["017i!QRϪH3=7+a$XƲw jJn{e&s#1x*r>b HxagѪM+y[׭E7"Fv;FJm9^yǜ ڧkXɍ*nLZ#E10[@I:U h vY%Q˕8᭿->*4s{v[;go$cz_1C ~(\:?1j{.?o6cA ϸm sJcWyiZ"NF,y{aWE9&3t8͢|;Y4ɢ y*H;zu/º5A4}Fv@?9ں5iА tJTYxo ԒW_ :ס\v6;7f: ~$~g曆יwwO86 {ybxGo~/ ^;iUفƑF:\w.׌awhHhhdA_֟U}=D#Ua j|ې(Yꏆ©ԙU8 ; #Tpr۵yZ;mkBrڮ-S0naK;?}خ]:E b|맩PVx v1/eeQVWϏn^Rrq~7Eo6nCVv0WjDye~EN&Wa8MzXI_jƄiĞ!裡.ͪ)yRvEM,ߠ=^#|n>'V3Mr2=Ve X=)3Ї  1ܓ^+7ٲo6y8 .7)ZPG7v+?Yhڥ@Ȃ@z퍀H ȯєŧ#4q%@6PCL'Z Ւ=r}"~{s$Ɍb*@@sbq0Eaʍo\e=)mc 0h[jȣm~1rϜ$r̭oTg/2JK I*;y_(TS_Q:LT"ZJ=ސ7ֲ"fƩ.d,ʑXlYp.*e˒r)N3oqĨZԭG/e*'LBɝ Ad]DQrE[UTUuToBC5 6stff#mW흶om~e˜1Aunu̯;PG]F uiNz/Jl, kĚLI@:.ɩ|noO VdCդzcb~"CM5 qIī̃W5@`x n- D-hGJЁ1a*3RpƂLiCʹÆZT灧/R[Qu8&xAEJ'jp+Ņ,!>%5-]5dt Iؽm@1j'!za2;<µ~ 7Q7v2>,]Vs_yK [ovƑ~YҀ&M!mo,Sd[nJo2Ul6j4 `bF?!^уur֕5e!{Kz('%d͐2zph$#fm6R4#r!\g;Yߍ-W~tϋټdE .p0 ρ.m^ >[,7\ og(OjdH\ L ^y"߫NN6*K.^|EA ]R4iת0"E~UK_i=$K1FaQ5V Y#4`ϙ\xa T(q[/FovSlce:eoenG6XC~~.^=86Bu V)1-Q _Iǡ99-/bz{QHgRU / ѝ46hoNs98ʜТG23 H, RPa^)S9"f9紶[N w^[{Sj|LDDSXBsElnѽ,0 eಪs2zX* j)z$Qҽy[2A:agH529EG %:נўw,s1eżfVbOQ+Fx(ڽxڡav5{cֳZvZlpU}D+M~MkHr \CU?eW_E]`7XFZ*aʡ1 Ryx^h1B._-a|}oXVm#| gծo?a .Bϳ)qdO3s{?K ̤w"Y5?MFz鬲93[bXPøɼN JR\\yTgg~[76MRs23 ”LM)Mzhs|o 2zz`j~΄1[rv3 q2R>?NBQ}#?_i&:>,,)hS'?Rc_ZyO/at=:1C=:ܞC&=0cdX.j7Zhסn, IdQxV%3TM7=S-%H^$vԧe/XZ/"Ňt*lTl0J:Cΐ淽wv4^Go޹m o\mmG ߓ2;Z=mܨ3]ضJud J>..@w:\0>{CݖZZUmIJ)۬va{ m==MOyAb)D?wsu9?,f$ 5d f޿ޖ/H}ǚ&4ns+q^J/iK3XįnqMd ٧fv\Z_e5U=YwnI6?z7U;8wK tRƻpv]}h5Dl6 :wSpBA餾w;*BIӐ#[hg)j u-2[c[ήZO0No&865'6-B}g]8 q RHҞK1-Y1Jvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004031376015136655267017721 0ustar rootrootJan 29 12:05:14 crc systemd[1]: Starting Kubernetes Kubelet... Jan 29 12:05:14 crc restorecon[4774]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:14 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 12:05:15 crc restorecon[4774]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 29 12:05:16 crc kubenswrapper[4993]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.916256 4993 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921633 4993 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921671 4993 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921686 4993 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921697 4993 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921710 4993 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921722 4993 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921733 4993 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921744 4993 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921786 4993 feature_gate.go:330] unrecognized feature gate: Example Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921797 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921808 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921841 4993 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921853 4993 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921863 4993 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921874 4993 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921886 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921897 4993 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921907 4993 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921918 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921929 4993 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921941 4993 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921954 4993 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921966 4993 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921981 4993 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.921995 4993 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922009 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922021 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922034 4993 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922046 4993 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922056 4993 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922067 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922079 4993 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922096 4993 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922111 4993 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922122 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922134 4993 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922145 4993 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922156 4993 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922167 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922179 4993 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922228 4993 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922239 4993 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922250 4993 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922261 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922272 4993 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922283 4993 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922294 4993 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922304 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922315 4993 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922326 4993 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922337 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922347 4993 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922364 4993 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922376 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922389 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922400 4993 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922412 4993 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922425 4993 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922439 4993 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922451 4993 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922462 4993 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922478 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922489 4993 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922500 4993 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922511 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922522 4993 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922533 4993 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922544 4993 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922555 4993 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922565 4993 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.922577 4993 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.923990 4993 flags.go:64] FLAG: --address="0.0.0.0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924025 4993 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924051 4993 flags.go:64] FLAG: --anonymous-auth="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924067 4993 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924084 4993 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924098 4993 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924114 4993 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924129 4993 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924143 4993 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924155 4993 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924170 4993 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924220 4993 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924235 4993 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924248 4993 flags.go:64] FLAG: --cgroup-root="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924260 4993 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924273 4993 flags.go:64] FLAG: --client-ca-file="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924286 4993 flags.go:64] FLAG: --cloud-config="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924299 4993 flags.go:64] FLAG: --cloud-provider="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924314 4993 flags.go:64] FLAG: --cluster-dns="[]" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924329 4993 flags.go:64] FLAG: --cluster-domain="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924341 4993 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924354 4993 flags.go:64] FLAG: --config-dir="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924367 4993 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924381 4993 flags.go:64] FLAG: --container-log-max-files="5" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924399 4993 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924411 4993 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924425 4993 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924438 4993 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924451 4993 flags.go:64] FLAG: --contention-profiling="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924464 4993 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924477 4993 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924490 4993 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924503 4993 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924519 4993 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924532 4993 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924545 4993 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924557 4993 flags.go:64] FLAG: --enable-load-reader="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924570 4993 flags.go:64] FLAG: --enable-server="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924583 4993 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924600 4993 flags.go:64] FLAG: --event-burst="100" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924615 4993 flags.go:64] FLAG: --event-qps="50" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924627 4993 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924640 4993 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924653 4993 flags.go:64] FLAG: --eviction-hard="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924668 4993 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924680 4993 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924693 4993 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924706 4993 flags.go:64] FLAG: --eviction-soft="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924718 4993 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924732 4993 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924746 4993 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924758 4993 flags.go:64] FLAG: --experimental-mounter-path="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924771 4993 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924783 4993 flags.go:64] FLAG: --fail-swap-on="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924796 4993 flags.go:64] FLAG: --feature-gates="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924811 4993 flags.go:64] FLAG: --file-check-frequency="20s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924824 4993 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924837 4993 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924850 4993 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924863 4993 flags.go:64] FLAG: --healthz-port="10248" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924876 4993 flags.go:64] FLAG: --help="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924889 4993 flags.go:64] FLAG: --hostname-override="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924901 4993 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924914 4993 flags.go:64] FLAG: --http-check-frequency="20s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924927 4993 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924939 4993 flags.go:64] FLAG: --image-credential-provider-config="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924950 4993 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924962 4993 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924974 4993 flags.go:64] FLAG: --image-service-endpoint="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924986 4993 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.924999 4993 flags.go:64] FLAG: --kube-api-burst="100" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925012 4993 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925026 4993 flags.go:64] FLAG: --kube-api-qps="50" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925038 4993 flags.go:64] FLAG: --kube-reserved="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925051 4993 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925063 4993 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925079 4993 flags.go:64] FLAG: --kubelet-cgroups="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925091 4993 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925104 4993 flags.go:64] FLAG: --lock-file="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925116 4993 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925129 4993 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925143 4993 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925165 4993 flags.go:64] FLAG: --log-json-split-stream="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925178 4993 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925227 4993 flags.go:64] FLAG: --log-text-split-stream="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925241 4993 flags.go:64] FLAG: --logging-format="text" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925254 4993 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925268 4993 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925280 4993 flags.go:64] FLAG: --manifest-url="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925293 4993 flags.go:64] FLAG: --manifest-url-header="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925310 4993 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925323 4993 flags.go:64] FLAG: --max-open-files="1000000" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925337 4993 flags.go:64] FLAG: --max-pods="110" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925350 4993 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925364 4993 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925377 4993 flags.go:64] FLAG: --memory-manager-policy="None" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925389 4993 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925403 4993 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925415 4993 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925428 4993 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925456 4993 flags.go:64] FLAG: --node-status-max-images="50" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925470 4993 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925482 4993 flags.go:64] FLAG: --oom-score-adj="-999" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925495 4993 flags.go:64] FLAG: --pod-cidr="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925507 4993 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925525 4993 flags.go:64] FLAG: --pod-manifest-path="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925538 4993 flags.go:64] FLAG: --pod-max-pids="-1" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925551 4993 flags.go:64] FLAG: --pods-per-core="0" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925564 4993 flags.go:64] FLAG: --port="10250" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925576 4993 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925590 4993 flags.go:64] FLAG: --provider-id="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925603 4993 flags.go:64] FLAG: --qos-reserved="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925619 4993 flags.go:64] FLAG: --read-only-port="10255" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925631 4993 flags.go:64] FLAG: --register-node="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925644 4993 flags.go:64] FLAG: --register-schedulable="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925656 4993 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925679 4993 flags.go:64] FLAG: --registry-burst="10" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925691 4993 flags.go:64] FLAG: --registry-qps="5" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925703 4993 flags.go:64] FLAG: --reserved-cpus="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925715 4993 flags.go:64] FLAG: --reserved-memory="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925730 4993 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925786 4993 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925799 4993 flags.go:64] FLAG: --rotate-certificates="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925812 4993 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925824 4993 flags.go:64] FLAG: --runonce="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925838 4993 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925851 4993 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925865 4993 flags.go:64] FLAG: --seccomp-default="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925878 4993 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925890 4993 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925903 4993 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925916 4993 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925930 4993 flags.go:64] FLAG: --storage-driver-password="root" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925945 4993 flags.go:64] FLAG: --storage-driver-secure="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925958 4993 flags.go:64] FLAG: --storage-driver-table="stats" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925972 4993 flags.go:64] FLAG: --storage-driver-user="root" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.925985 4993 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926001 4993 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926014 4993 flags.go:64] FLAG: --system-cgroups="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926027 4993 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926047 4993 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926060 4993 flags.go:64] FLAG: --tls-cert-file="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926073 4993 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926087 4993 flags.go:64] FLAG: --tls-min-version="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926100 4993 flags.go:64] FLAG: --tls-private-key-file="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926111 4993 flags.go:64] FLAG: --topology-manager-policy="none" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926124 4993 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926137 4993 flags.go:64] FLAG: --topology-manager-scope="container" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926152 4993 flags.go:64] FLAG: --v="2" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926168 4993 flags.go:64] FLAG: --version="false" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926219 4993 flags.go:64] FLAG: --vmodule="" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926235 4993 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.926250 4993 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926575 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926594 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926608 4993 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926620 4993 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926631 4993 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926643 4993 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926654 4993 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926669 4993 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926683 4993 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926695 4993 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926707 4993 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926718 4993 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926731 4993 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926744 4993 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926755 4993 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926768 4993 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926781 4993 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926792 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926804 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926816 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926828 4993 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926840 4993 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926851 4993 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926862 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926874 4993 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926886 4993 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926900 4993 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926911 4993 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926922 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926933 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926945 4993 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926958 4993 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926970 4993 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926980 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.926991 4993 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927002 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927014 4993 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927024 4993 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927036 4993 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927046 4993 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927057 4993 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927072 4993 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927084 4993 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927096 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927106 4993 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927117 4993 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927133 4993 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927145 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927157 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927171 4993 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927221 4993 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927238 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927251 4993 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927263 4993 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927278 4993 feature_gate.go:330] unrecognized feature gate: Example Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927290 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927301 4993 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927313 4993 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927325 4993 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927529 4993 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927540 4993 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927550 4993 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927561 4993 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927572 4993 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927582 4993 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927592 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927602 4993 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927615 4993 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927625 4993 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927635 4993 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.927646 4993 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.927662 4993 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.948277 4993 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.948335 4993 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948538 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948553 4993 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948564 4993 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948573 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948584 4993 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948593 4993 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948602 4993 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948612 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948624 4993 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948641 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948654 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948665 4993 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948678 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948689 4993 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948700 4993 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948712 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948723 4993 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948736 4993 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948750 4993 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948762 4993 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948776 4993 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948786 4993 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948797 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948809 4993 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948820 4993 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948832 4993 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948843 4993 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948860 4993 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948873 4993 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948885 4993 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948895 4993 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948904 4993 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948913 4993 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948923 4993 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948933 4993 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948942 4993 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948951 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948960 4993 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948969 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948978 4993 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948989 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.948999 4993 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949007 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949017 4993 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949026 4993 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949036 4993 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949046 4993 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949054 4993 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949064 4993 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949073 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949083 4993 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949098 4993 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949110 4993 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949121 4993 feature_gate.go:330] unrecognized feature gate: Example Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949132 4993 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949142 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949153 4993 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949164 4993 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949175 4993 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949222 4993 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949233 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949243 4993 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949255 4993 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949266 4993 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949276 4993 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949287 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949297 4993 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949309 4993 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949320 4993 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949330 4993 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949341 4993 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.949358 4993 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949661 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949685 4993 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949698 4993 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949710 4993 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949754 4993 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949764 4993 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949772 4993 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949781 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949789 4993 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949798 4993 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949809 4993 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949820 4993 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949828 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949837 4993 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949846 4993 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949855 4993 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949863 4993 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949871 4993 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949879 4993 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949888 4993 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949896 4993 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949905 4993 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949913 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949922 4993 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949930 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949941 4993 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949953 4993 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949964 4993 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949974 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949983 4993 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.949992 4993 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950001 4993 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950009 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950017 4993 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950026 4993 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950034 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950043 4993 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950053 4993 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950062 4993 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950070 4993 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950079 4993 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950088 4993 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950099 4993 feature_gate.go:330] unrecognized feature gate: Example Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950110 4993 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950121 4993 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950135 4993 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950149 4993 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950162 4993 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950173 4993 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950223 4993 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950237 4993 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950249 4993 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950260 4993 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950274 4993 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950285 4993 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950296 4993 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950304 4993 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950313 4993 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950321 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950330 4993 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950338 4993 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950347 4993 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950356 4993 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950364 4993 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950372 4993 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950382 4993 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950391 4993 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950399 4993 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950407 4993 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950416 4993 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 12:05:16 crc kubenswrapper[4993]: W0129 12:05:16.950424 4993 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.950437 4993 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.950736 4993 server.go:940] "Client rotation is on, will bootstrap in background" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.958430 4993 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.958586 4993 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.962062 4993 server.go:997] "Starting client certificate rotation" Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.962129 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.962445 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-09 21:41:41.059790383 +0000 UTC Jan 29 12:05:16 crc kubenswrapper[4993]: I0129 12:05:16.962573 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.015180 4993 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.017911 4993 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.018957 4993 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.036492 4993 log.go:25] "Validated CRI v1 runtime API" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.090539 4993 log.go:25] "Validated CRI v1 image API" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.092157 4993 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.096447 4993 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-29-12-00-09-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.096483 4993 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.113841 4993 manager.go:217] Machine: {Timestamp:2026-01-29 12:05:17.111289754 +0000 UTC m=+1.126419920 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6d0a6411-7fd6-4d96-a862-4c3278683d9f BootID:2ddad0c9-9bb7-4482-acb5-3fbc806f4a38 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7e:fa:7d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7e:fa:7d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0c:9c:94 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:29:a0:b1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1e:83:39 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:47:38:df Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:1e:9a:0d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5a:d4:59:0b:02:27 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:06:72:82:25:3e:54 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.114114 4993 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.114272 4993 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.116692 4993 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.116953 4993 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.116998 4993 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.117254 4993 topology_manager.go:138] "Creating topology manager with none policy" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.117266 4993 container_manager_linux.go:303] "Creating device plugin manager" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.117906 4993 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.117948 4993 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.118209 4993 state_mem.go:36] "Initialized new in-memory state store" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.118305 4993 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.124463 4993 kubelet.go:418] "Attempting to sync node with API server" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.124492 4993 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.124532 4993 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.124553 4993 kubelet.go:324] "Adding apiserver pod source" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.124567 4993 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.130251 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.130335 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.130330 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.130443 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.131627 4993 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.132730 4993 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.134953 4993 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.136933 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.136987 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137003 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137017 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137038 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137053 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137066 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137088 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137104 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137117 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137153 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.137166 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.139568 4993 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.140269 4993 server.go:1280] "Started kubelet" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.140410 4993 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.140605 4993 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.141067 4993 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.141284 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:17 crc systemd[1]: Started Kubernetes Kubelet. Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.142670 4993 server.go:460] "Adding debug handlers to kubelet server" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.143734 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.143770 4993 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.143906 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 18:25:06.496172987 +0000 UTC Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.144034 4993 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.145455 4993 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.145532 4993 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.145553 4993 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.145520 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.146581 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.146689 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.148275 4993 factory.go:55] Registering systemd factory Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.148303 4993 factory.go:221] Registration of the systemd container factory successfully Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.149135 4993 factory.go:153] Registering CRI-O factory Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.149247 4993 factory.go:221] Registration of the crio container factory successfully Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.149363 4993 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.149453 4993 factory.go:103] Registering Raw factory Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.149575 4993 manager.go:1196] Started watching for new ooms in manager Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.148424 4993 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188f3225f24d312c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 12:05:17.140218156 +0000 UTC m=+1.155348322,LastTimestamp:2026-01-29 12:05:17.140218156 +0000 UTC m=+1.155348322,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.154607 4993 manager.go:319] Starting recovery of all containers Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155507 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155559 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155572 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155583 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155594 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155604 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155619 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155631 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155644 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155654 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155664 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155674 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155684 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155696 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155708 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155718 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155727 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155736 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155746 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155757 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155769 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155778 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155788 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155797 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155808 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155819 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155833 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155845 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155860 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155872 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155883 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155896 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155906 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155915 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155926 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155937 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155947 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155956 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155966 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.155979 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156001 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156018 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156031 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156047 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156059 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156071 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156092 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156103 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156115 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156126 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156137 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156148 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156165 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156176 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156210 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156221 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156233 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156244 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156254 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156263 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156274 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156284 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156296 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156307 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156318 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156330 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156341 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156352 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156361 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156371 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156381 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156393 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156403 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156413 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156425 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156434 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156447 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156477 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156489 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156499 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156510 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156520 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156529 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156539 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156552 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156561 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156571 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156581 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156590 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.156601 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158329 4993 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158368 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158389 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158401 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158421 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158432 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158443 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158456 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158472 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158484 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158494 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158504 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158514 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158527 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158538 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158556 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158569 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158581 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158592 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158605 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158618 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158631 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158642 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158652 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158663 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158673 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158684 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158695 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158705 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158715 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158726 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158735 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158745 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158756 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158766 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158776 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158787 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158798 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158808 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158819 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158828 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158842 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158853 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158866 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158879 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158890 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158901 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158911 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158922 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158943 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158955 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158967 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158980 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.158990 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159000 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159010 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159021 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159033 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159044 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159055 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159065 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159075 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159085 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159095 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159106 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159116 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159126 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159135 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159146 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159156 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159166 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159176 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159211 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159221 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159231 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159242 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159254 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159263 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159274 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159283 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159293 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159306 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159317 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159327 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159338 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159348 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159364 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159374 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159383 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159394 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159403 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159413 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159423 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159433 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159448 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159458 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159468 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159480 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159490 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159500 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159512 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159524 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159537 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159559 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159575 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159587 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159597 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159609 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159621 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159631 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159651 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159662 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159672 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159683 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159693 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159703 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159714 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159726 4993 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159735 4993 reconstruct.go:97] "Volume reconstruction finished" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.159742 4993 reconciler.go:26] "Reconciler: start to sync state" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.170248 4993 manager.go:324] Recovery completed Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.182627 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.184367 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.184400 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.184410 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.185093 4993 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.185112 4993 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.185134 4993 state_mem.go:36] "Initialized new in-memory state store" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.187171 4993 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.189152 4993 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.189209 4993 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.189254 4993 kubelet.go:2335] "Starting kubelet main sync loop" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.189304 4993 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.191249 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.191312 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.206456 4993 policy_none.go:49] "None policy: Start" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.207506 4993 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.207536 4993 state_mem.go:35] "Initializing new in-memory state store" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.244507 4993 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.260365 4993 manager.go:334] "Starting Device Plugin manager" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.260427 4993 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.260440 4993 server.go:79] "Starting device plugin registration server" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.260986 4993 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.261011 4993 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.261283 4993 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.261394 4993 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.261405 4993 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.272868 4993 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.289999 4993 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.290132 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291483 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291536 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291675 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291797 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.291836 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295421 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295539 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295421 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295591 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.295769 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.296028 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.296137 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.296789 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.296842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.296854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297030 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297214 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297252 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297649 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.297659 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299036 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299250 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299270 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299281 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299254 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299344 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.299376 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.301251 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.301291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.301302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.301501 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.301547 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.302736 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.302764 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.302776 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.311012 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.311081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.311093 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.346574 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361311 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361410 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361450 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361481 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361501 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361525 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361546 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361569 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361599 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361620 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361637 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361652 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361666 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361681 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361767 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.361816 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.362268 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.362299 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.362309 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.362329 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.362718 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.462962 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463020 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463044 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463064 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463083 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463102 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463110 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463156 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463173 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463125 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463228 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463214 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463226 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463223 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463259 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463495 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463538 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463551 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463562 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463607 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463276 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463628 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463632 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463657 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463679 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463702 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463724 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463732 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463770 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.463849 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.563599 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.565883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.565944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.565975 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.566016 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.566546 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.634044 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.655145 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.664426 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.682707 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.689365 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.694638 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-43dcd39dfcaa4b5739f8478b18369f9987ae38935b15d440dac8bacefe07cd48 WatchSource:0}: Error finding container 43dcd39dfcaa4b5739f8478b18369f9987ae38935b15d440dac8bacefe07cd48: Status 404 returned error can't find the container with id 43dcd39dfcaa4b5739f8478b18369f9987ae38935b15d440dac8bacefe07cd48 Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.707538 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-290a39b605003333828d3b45a4448cd8937dd3b4df3c1ff0ea950d49887f4013 WatchSource:0}: Error finding container 290a39b605003333828d3b45a4448cd8937dd3b4df3c1ff0ea950d49887f4013: Status 404 returned error can't find the container with id 290a39b605003333828d3b45a4448cd8937dd3b4df3c1ff0ea950d49887f4013 Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.711602 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7d6ccde719256f3b5cc66adeff0b33d8fd834d823088b4261a8c9806b93de355 WatchSource:0}: Error finding container 7d6ccde719256f3b5cc66adeff0b33d8fd834d823088b4261a8c9806b93de355: Status 404 returned error can't find the container with id 7d6ccde719256f3b5cc66adeff0b33d8fd834d823088b4261a8c9806b93de355 Jan 29 12:05:17 crc kubenswrapper[4993]: W0129 12:05:17.716745 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9b3e6230e6440b4afacdba0ef2cbfb47c6f9989490dc800d5f1ea95ab087243e WatchSource:0}: Error finding container 9b3e6230e6440b4afacdba0ef2cbfb47c6f9989490dc800d5f1ea95ab087243e: Status 404 returned error can't find the container with id 9b3e6230e6440b4afacdba0ef2cbfb47c6f9989490dc800d5f1ea95ab087243e Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.747382 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.966861 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.968877 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.968928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.968942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:17 crc kubenswrapper[4993]: I0129 12:05:17.968981 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:17 crc kubenswrapper[4993]: E0129 12:05:17.969623 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 12:05:18 crc kubenswrapper[4993]: W0129 12:05:18.011350 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.011415 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.142779 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.144863 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 08:16:11.260372528 +0000 UTC Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.196474 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"969325e965ebc37f8b217af8016f564a4bd2b029550e822293a77d22441a6a7b"} Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.197985 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"43dcd39dfcaa4b5739f8478b18369f9987ae38935b15d440dac8bacefe07cd48"} Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.199149 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9b3e6230e6440b4afacdba0ef2cbfb47c6f9989490dc800d5f1ea95ab087243e"} Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.200398 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d6ccde719256f3b5cc66adeff0b33d8fd834d823088b4261a8c9806b93de355"} Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.201492 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"290a39b605003333828d3b45a4448cd8937dd3b4df3c1ff0ea950d49887f4013"} Jan 29 12:05:18 crc kubenswrapper[4993]: W0129 12:05:18.214281 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.214366 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:18 crc kubenswrapper[4993]: W0129 12:05:18.356145 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.356692 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.548558 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Jan 29 12:05:18 crc kubenswrapper[4993]: W0129 12:05:18.763718 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.763876 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.770735 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.773485 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.773527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.773540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:18 crc kubenswrapper[4993]: I0129 12:05:18.773574 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:18 crc kubenswrapper[4993]: E0129 12:05:18.773921 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.142753 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.145849 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 17:05:33.088032967 +0000 UTC Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.176317 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 12:05:19 crc kubenswrapper[4993]: E0129 12:05:19.178031 4993 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.207969 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208013 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208018 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208023 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208100 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208813 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.208831 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.209528 4993 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6" exitCode=0 Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.209600 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.209617 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.210412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.210481 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.210508 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.211007 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa" exitCode=0 Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.211072 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.211240 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212311 4993 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2" exitCode=0 Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212895 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212373 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.212462 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214023 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214050 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214877 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214910 4993 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca" exitCode=0 Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.214941 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca"} Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.215000 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.215920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.215953 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.215964 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.217244 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.217287 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:19 crc kubenswrapper[4993]: I0129 12:05:19.217302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.142621 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.146807 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 20:44:41.629983946 +0000 UTC Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.149591 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.220568 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.220615 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.220628 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.220632 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.221737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.221767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.221778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.223638 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.223671 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.223688 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.225556 4993 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc" exitCode=0 Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.225623 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.225786 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.227245 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.227312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.227331 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.228372 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd"} Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.228404 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.228419 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229565 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229597 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229688 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229668 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.229705 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: W0129 12:05:20.344375 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.344490 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:20 crc kubenswrapper[4993]: W0129 12:05:20.344788 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.344872 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:20 crc kubenswrapper[4993]: W0129 12:05:20.359219 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.359308 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.374699 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.376556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.376593 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.376602 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:20 crc kubenswrapper[4993]: I0129 12:05:20.376623 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.377079 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 12:05:20 crc kubenswrapper[4993]: W0129 12:05:20.744015 4993 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:20 crc kubenswrapper[4993]: E0129 12:05:20.744092 4993 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.142666 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.146950 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 01:13:46.756218231 +0000 UTC Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.234554 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a"} Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.234624 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4"} Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.234689 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.236354 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.236401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.236412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237628 4993 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7" exitCode=0 Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237694 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237746 4993 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237782 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237732 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7"} Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.237881 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238832 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238861 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238931 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.238972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.239286 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.239316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:21 crc kubenswrapper[4993]: I0129 12:05:21.239329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.147144 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 19:45:40.270968881 +0000 UTC Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.242988 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dda7a5c8db3f88f861b71392d31f5df467c7de45e03192201438899b6847799e"} Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.243043 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5db7fb82b3a528a53375b9213a3af252b8e9fd80d23731c1660d08ce94e25749"} Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.243069 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9f34aad679fef25c08a240387412b0d423afcb66b6b50338926a355e1d2c2cc2"} Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.243088 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fed706c97954479b34d7cef65cc6d0749bf8f064e8023b643899d3d84276da8f"} Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.243094 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.243125 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.244419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.244451 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:22 crc kubenswrapper[4993]: I0129 12:05:22.244459 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.148153 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 14:30:26.094372885 +0000 UTC Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.248626 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2dd3a2b8ef581a30ae5ad4678d5e2dcda7c53d96d219baa1363d5198baae9731"} Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.248689 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.248723 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251903 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.251960 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.478090 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.578018 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.579605 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.579657 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.579673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.579704 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:23 crc kubenswrapper[4993]: I0129 12:05:23.914955 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.045110 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.149249 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:29:12.887814418 +0000 UTC Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.251293 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.251348 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252306 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252347 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252360 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.252415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.834973 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.835475 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.836617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.836669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.836681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:24 crc kubenswrapper[4993]: I0129 12:05:24.840112 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.150305 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 10:32:36.265751225 +0000 UTC Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.253547 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.253615 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.253628 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.254554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.254648 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.254661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.255470 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.255501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.255511 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.488906 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.489234 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.490834 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.490884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.490901 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:25 crc kubenswrapper[4993]: I0129 12:05:25.554853 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.150905 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 07:05:55.798541793 +0000 UTC Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.256465 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.257673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.257724 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.257741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.939880 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.940067 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.941150 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.941179 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:26 crc kubenswrapper[4993]: I0129 12:05:26.941211 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:27 crc kubenswrapper[4993]: I0129 12:05:27.152024 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 18:59:00.87254948 +0000 UTC Jan 29 12:05:27 crc kubenswrapper[4993]: I0129 12:05:27.260089 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:27 crc kubenswrapper[4993]: I0129 12:05:27.261380 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:27 crc kubenswrapper[4993]: I0129 12:05:27.261431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:27 crc kubenswrapper[4993]: I0129 12:05:27.261451 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:27 crc kubenswrapper[4993]: E0129 12:05:27.273931 4993 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.152481 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 05:32:19.485874109 +0000 UTC Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.248707 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.263793 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.265805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.266763 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.266788 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:28 crc kubenswrapper[4993]: I0129 12:05:28.270510 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:29 crc kubenswrapper[4993]: I0129 12:05:29.153144 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 08:21:29.021389843 +0000 UTC Jan 29 12:05:29 crc kubenswrapper[4993]: I0129 12:05:29.266237 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:29 crc kubenswrapper[4993]: I0129 12:05:29.267814 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:29 crc kubenswrapper[4993]: I0129 12:05:29.267906 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:29 crc kubenswrapper[4993]: I0129 12:05:29.267926 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:30 crc kubenswrapper[4993]: I0129 12:05:30.153619 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 17:03:04.295529459 +0000 UTC Jan 29 12:05:31 crc kubenswrapper[4993]: I0129 12:05:31.154826 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 23:36:25.476580624 +0000 UTC Jan 29 12:05:31 crc kubenswrapper[4993]: I0129 12:05:31.248807 4993 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 12:05:31 crc kubenswrapper[4993]: I0129 12:05:31.248885 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 12:05:31 crc kubenswrapper[4993]: I0129 12:05:31.675797 4993 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44448->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 29 12:05:31 crc kubenswrapper[4993]: I0129 12:05:31.675862 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44448->192.168.126.11:17697: read: connection reset by peer" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.144000 4993 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.155511 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 15:50:11.486291907 +0000 UTC Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.273459 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.275642 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a" exitCode=255 Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.276346 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a"} Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.276615 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.277670 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.277700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.277711 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.278177 4993 scope.go:117] "RemoveContainer" containerID="15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.294378 4993 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.294441 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.315937 4993 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.316007 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.506994 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.507144 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.508089 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.508328 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.508412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:32 crc kubenswrapper[4993]: I0129 12:05:32.587429 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.156498 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 09:22:45.385083315 +0000 UTC Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.287282 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.289347 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0"} Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.289428 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.289518 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290290 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290389 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290420 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.290431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:33 crc kubenswrapper[4993]: I0129 12:05:33.314433 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.049567 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.157124 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 04:24:23.889630665 +0000 UTC Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.291863 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.291925 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.291882 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293242 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293286 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293471 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.293501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:34 crc kubenswrapper[4993]: I0129 12:05:34.304641 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:35 crc kubenswrapper[4993]: I0129 12:05:35.157912 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 17:27:02.056100101 +0000 UTC Jan 29 12:05:35 crc kubenswrapper[4993]: I0129 12:05:35.294469 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:35 crc kubenswrapper[4993]: I0129 12:05:35.295399 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:35 crc kubenswrapper[4993]: I0129 12:05:35.295661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:35 crc kubenswrapper[4993]: I0129 12:05:35.295841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:36 crc kubenswrapper[4993]: I0129 12:05:36.158604 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 14:45:41.791096709 +0000 UTC Jan 29 12:05:36 crc kubenswrapper[4993]: I0129 12:05:36.295869 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:36 crc kubenswrapper[4993]: I0129 12:05:36.296597 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:36 crc kubenswrapper[4993]: I0129 12:05:36.296626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:36 crc kubenswrapper[4993]: I0129 12:05:36.296635 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.159432 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 23:43:38.498533971 +0000 UTC Jan 29 12:05:37 crc kubenswrapper[4993]: E0129 12:05:37.274808 4993 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 12:05:37 crc kubenswrapper[4993]: E0129 12:05:37.297599 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.299623 4993 trace.go:236] Trace[805675872]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 12:05:24.944) (total time: 12355ms): Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[805675872]: ---"Objects listed" error: 12355ms (12:05:37.299) Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[805675872]: [12.355210766s] [12.355210766s] END Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.299668 4993 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.300732 4993 trace.go:236] Trace[1091408589]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 12:05:24.053) (total time: 13247ms): Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[1091408589]: ---"Objects listed" error: 13247ms (12:05:37.300) Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[1091408589]: [13.247094793s] [13.247094793s] END Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.300915 4993 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 12:05:37 crc kubenswrapper[4993]: E0129 12:05:37.301224 4993 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.301951 4993 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.301986 4993 trace.go:236] Trace[641020003]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 12:05:26.695) (total time: 10606ms): Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[641020003]: ---"Objects listed" error: 10606ms (12:05:37.301) Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[641020003]: [10.606845222s] [10.606845222s] END Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.302010 4993 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.303779 4993 trace.go:236] Trace[1133755950]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 12:05:26.604) (total time: 10699ms): Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[1133755950]: ---"Objects listed" error: 10699ms (12:05:37.303) Jan 29 12:05:37 crc kubenswrapper[4993]: Trace[1133755950]: [10.699181538s] [10.699181538s] END Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.303818 4993 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.304733 4993 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.318913 4993 csr.go:261] certificate signing request csr-lghpr is approved, waiting to be issued Jan 29 12:05:37 crc kubenswrapper[4993]: I0129 12:05:37.327724 4993 csr.go:257] certificate signing request csr-lghpr is issued Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.137060 4993 apiserver.go:52] "Watching apiserver" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.159735 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 12:18:27.296338669 +0000 UTC Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.164222 4993 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.164526 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.165024 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.165101 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.165126 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.165249 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.165303 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.165798 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.165812 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.166314 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.166352 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.167687 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.167921 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.168233 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.168375 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.169227 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.169911 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.171011 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.171697 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.171872 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.196492 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.211601 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.224761 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.234480 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.243079 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.246454 4993 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.253754 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.256093 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.261999 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.262317 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.271759 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.279627 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322145 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322479 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322532 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322564 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322597 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322632 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322663 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322700 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322731 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322759 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322791 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322823 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322853 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322884 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322912 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322941 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322966 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.322995 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323027 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323059 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323084 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323118 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323147 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323178 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323224 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323255 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323288 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323316 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323345 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323377 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323401 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323435 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323466 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323495 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323520 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323555 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323596 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323627 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323658 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323685 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323721 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323754 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323781 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323817 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323848 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323872 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323904 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323942 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323971 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.323997 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324025 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324053 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324078 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324107 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324135 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324229 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324256 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324284 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324315 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324342 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324374 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324404 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324430 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324459 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324487 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324518 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324542 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324569 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324600 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324617 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324623 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324848 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324892 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324957 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324970 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325036 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325065 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325126 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325603 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325695 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325883 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325910 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325929 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325961 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.325984 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326115 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326181 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326246 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326318 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326323 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326645 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326685 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326691 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326707 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326970 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.326930 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327022 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327121 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327035 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327174 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327237 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327285 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327521 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.324636 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327590 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327629 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327662 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327697 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327730 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327779 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327811 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327840 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327869 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327894 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327922 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327956 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327984 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328013 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328044 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328070 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328098 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328158 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328203 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328225 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328245 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328265 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328281 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328301 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328320 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328339 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328358 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328377 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328397 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328417 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328437 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328457 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328476 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328497 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328518 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328538 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328557 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328580 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328603 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328622 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328641 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328660 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328679 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328701 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328724 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328760 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328790 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328818 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328846 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328872 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328894 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328915 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328934 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328953 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328973 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329180 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329219 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329242 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329264 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329291 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329308 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329329 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329349 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329366 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329385 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329406 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329422 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329442 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329464 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329493 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329516 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329538 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329558 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329574 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329594 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329614 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329632 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329652 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329671 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329695 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329713 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329734 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329753 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329774 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329802 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329835 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329862 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329948 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329977 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330004 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330026 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330053 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330160 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330215 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330245 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330290 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330314 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330341 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330371 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330398 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330429 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330465 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330490 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330532 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330561 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330587 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327724 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327860 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327870 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.327912 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328098 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328218 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328240 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330697 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328407 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328528 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328553 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328692 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328694 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328758 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328914 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.328968 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329176 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329352 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329401 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329415 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329497 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329517 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329522 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329645 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.329863 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330066 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330070 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330102 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330521 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330628 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331130 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331218 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-29 12:00:37 +0000 UTC, rotation deadline is 2026-12-23 05:07:49.271958683 +0000 UTC Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331176 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331257 4993 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7865h2m10.940703528s for next certificate rotation Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331326 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331555 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331579 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331712 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331722 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.331831 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332058 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332059 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332177 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332361 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332371 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332478 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332687 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332699 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332751 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332775 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332774 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.332939 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333074 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333167 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333179 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333230 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333246 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333287 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333509 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333634 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333773 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333806 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.333844 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.335809 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.335844 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336210 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336259 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336338 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336468 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336548 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336640 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336875 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.336887 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337326 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337332 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337436 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337688 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337748 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337863 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.337924 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.338277 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.338728 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.338878 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339154 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339199 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339472 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339457 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339531 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339544 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.330646 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339676 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339707 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339731 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339752 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339774 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339793 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339810 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.340480 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.340500 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.341930 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342205 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342235 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342321 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342325 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342461 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342623 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342824 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342826 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342852 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.339832 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342918 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342947 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.342972 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343158 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343253 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343270 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343334 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343370 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343371 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.343482 4993 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343623 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343659 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343777 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343810 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343832 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343863 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343883 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.343903 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344110 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344170 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344271 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344321 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344456 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.344849 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.345444 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.346318 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:38.846302775 +0000 UTC m=+22.861432901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346357 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346398 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.345329 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346485 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346505 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346522 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346541 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346559 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346577 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346593 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346610 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346628 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346629 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346645 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346664 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346681 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346697 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346742 4993 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346753 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346764 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346773 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346781 4993 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346791 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346799 4993 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346807 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346816 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346824 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346834 4993 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346843 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346852 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346860 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346869 4993 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346878 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346887 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346896 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346904 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346912 4993 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346922 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346930 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346938 4993 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346946 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346955 4993 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346963 4993 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346972 4993 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346981 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346989 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346997 4993 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347005 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347018 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347033 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347044 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347056 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347067 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347079 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347090 4993 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347103 4993 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347113 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347122 4993 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347131 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347140 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347148 4993 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347157 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347166 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347174 4993 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347182 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347205 4993 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347214 4993 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347235 4993 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347244 4993 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347252 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347260 4993 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347269 4993 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347277 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347286 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347295 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347304 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347314 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347323 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347332 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347340 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347354 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347363 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347372 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347384 4993 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347393 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347401 4993 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347409 4993 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347418 4993 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347450 4993 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347462 4993 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347474 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347485 4993 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347496 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347508 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347517 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347528 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347540 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347550 4993 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347559 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347568 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347577 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347585 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347595 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347604 4993 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347614 4993 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347626 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347638 4993 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347649 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347661 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347672 4993 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347684 4993 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347695 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347706 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347718 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347730 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347744 4993 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347756 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347767 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347778 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347788 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347798 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347809 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347820 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347832 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347843 4993 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347854 4993 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347865 4993 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347877 4993 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347889 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347900 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347911 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347923 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347934 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347945 4993 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347956 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347965 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347975 4993 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347987 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347999 4993 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348010 4993 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348021 4993 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348033 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348044 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348072 4993 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348084 4993 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348095 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348106 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348118 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348129 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348140 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348151 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348160 4993 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348171 4993 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348225 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348239 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348251 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348262 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348274 4993 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348285 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348297 4993 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348308 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348319 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348329 4993 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348340 4993 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347014 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347045 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346119 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347307 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.347481 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348051 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.348464 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.348780 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.348820 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:38.848809961 +0000 UTC m=+22.863940127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.349286 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.346089 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.349699 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.349789 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.350314 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.350527 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.350550 4993 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.350736 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.350959 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.351368 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.351499 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.351651 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.351966 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.352317 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.352596 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.352844 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.352947 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.353177 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.353384 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.353511 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.353666 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.354375 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.354733 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.355040 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.355287 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.355509 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.355684 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.356263 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.356232 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.356804 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.357099 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.357423 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.357866 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.358169 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.358207 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.358294 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:05:38.858272753 +0000 UTC m=+22.873402989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.358515 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.358583 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.359108 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.362491 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.362873 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.363685 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.363721 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.364457 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.364479 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.364510 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.364574 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:38.86455193 +0000 UTC m=+22.879682046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.365900 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.366422 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.366467 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.372459 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.377661 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.381929 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.381967 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.381982 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.382043 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:38.882017664 +0000 UTC m=+22.897147790 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.388809 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.394465 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.398734 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.403911 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.414703 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449258 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449346 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449374 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449408 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449424 4993 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449437 4993 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449447 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449458 4993 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449470 4993 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449481 4993 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449491 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449504 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449552 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449565 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449576 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449587 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449599 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449610 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449521 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449621 4993 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449696 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449712 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449740 4993 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449758 4993 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449770 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449781 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449791 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449802 4993 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449813 4993 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449823 4993 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449836 4993 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449846 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449857 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449868 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449880 4993 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449891 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449902 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449915 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449927 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449939 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449950 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449962 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449976 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449987 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.449997 4993 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450009 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450021 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450032 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450044 4993 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450056 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.450067 4993 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.462804 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-bskq4"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.463135 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.463572 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-clvrz"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.464080 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.464755 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-b9gn8"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.464998 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467294 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467599 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467616 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467651 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467899 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.467951 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.468106 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.468211 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.468302 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.468400 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.468333 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.469176 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.470410 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.479500 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.482677 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.489503 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.500203 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.502719 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.514560 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: W0129 12:05:38.519626 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-33312ff181791b1819893a1d1cfd5060d78cff3cf0996930db8326b0096ec142 WatchSource:0}: Error finding container 33312ff181791b1819893a1d1cfd5060d78cff3cf0996930db8326b0096ec142: Status 404 returned error can't find the container with id 33312ff181791b1819893a1d1cfd5060d78cff3cf0996930db8326b0096ec142 Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.524418 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.540122 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553426 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-hosts-file\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553495 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1df6750-623d-4dda-83cd-dec067f8fc1a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553520 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-os-release\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553539 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlgfb\" (UniqueName: \"kubernetes.io/projected/b1df6750-623d-4dda-83cd-dec067f8fc1a-kube-api-access-zlgfb\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553560 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-multus\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553579 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-conf-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553610 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1df6750-623d-4dda-83cd-dec067f8fc1a-proxy-tls\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553630 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-socket-dir-parent\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553648 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-netns\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553706 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-daemon-config\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553735 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-multus-certs\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553764 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-hostroot\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553790 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qljfq\" (UniqueName: \"kubernetes.io/projected/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-kube-api-access-qljfq\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553812 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b1df6750-623d-4dda-83cd-dec067f8fc1a-rootfs\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553835 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cni-binary-copy\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553858 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553883 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-bin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553923 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-system-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553954 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cnibin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553981 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-k8s-cni-cncf-io\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.553998 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vfrd\" (UniqueName: \"kubernetes.io/projected/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-kube-api-access-9vfrd\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.554033 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-kubelet\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.554050 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-etc-kubernetes\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.559413 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.570425 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.582473 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.593520 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.603763 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.616098 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.624758 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.634457 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.645293 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654333 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b1df6750-623d-4dda-83cd-dec067f8fc1a-rootfs\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654379 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cni-binary-copy\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654403 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-hostroot\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654423 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qljfq\" (UniqueName: \"kubernetes.io/projected/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-kube-api-access-qljfq\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654445 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654467 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-bin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654499 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-k8s-cni-cncf-io\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654558 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vfrd\" (UniqueName: \"kubernetes.io/projected/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-kube-api-access-9vfrd\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654635 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-system-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654659 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cnibin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654683 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-kubelet\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654705 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-etc-kubernetes\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654844 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654945 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655040 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b1df6750-623d-4dda-83cd-dec067f8fc1a-rootfs\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655092 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-hostroot\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655135 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-k8s-cni-cncf-io\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655095 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-bin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655172 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cnibin\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655235 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-kubelet\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655261 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-etc-kubernetes\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655303 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-system-cni-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.654729 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1df6750-623d-4dda-83cd-dec067f8fc1a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655672 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-os-release\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655693 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-hosts-file\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655710 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlgfb\" (UniqueName: \"kubernetes.io/projected/b1df6750-623d-4dda-83cd-dec067f8fc1a-kube-api-access-zlgfb\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655726 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-multus\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655743 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-conf-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655757 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-socket-dir-parent\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655772 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-netns\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655791 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-cni-binary-copy\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655805 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1df6750-623d-4dda-83cd-dec067f8fc1a-proxy-tls\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.655873 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-daemon-config\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656042 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-os-release\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656069 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-hosts-file\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656098 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-conf-dir\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656121 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-var-lib-cni-multus\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656156 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-socket-dir-parent\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656220 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-multus-certs\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656333 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-multus-certs\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656223 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-host-run-netns\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.656896 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-multus-daemon-config\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.659410 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b1df6750-623d-4dda-83cd-dec067f8fc1a-mcd-auth-proxy-config\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.659706 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b1df6750-623d-4dda-83cd-dec067f8fc1a-proxy-tls\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.668122 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.670945 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlgfb\" (UniqueName: \"kubernetes.io/projected/b1df6750-623d-4dda-83cd-dec067f8fc1a-kube-api-access-zlgfb\") pod \"machine-config-daemon-bskq4\" (UID: \"b1df6750-623d-4dda-83cd-dec067f8fc1a\") " pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.671422 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vfrd\" (UniqueName: \"kubernetes.io/projected/4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8-kube-api-access-9vfrd\") pod \"node-resolver-clvrz\" (UID: \"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\") " pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.672894 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qljfq\" (UniqueName: \"kubernetes.io/projected/d22ea793-b0f9-4b2c-8ea7-ed61728d900f-kube-api-access-qljfq\") pod \"multus-b9gn8\" (UID: \"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\") " pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.677638 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.687390 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.776228 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.786214 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-clvrz" Jan 29 12:05:38 crc kubenswrapper[4993]: W0129 12:05:38.788347 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1df6750_623d_4dda_83cd_dec067f8fc1a.slice/crio-d260aab98481de9ce91264fba0b4bcff3062b11c51127d1582d69e54415c4cfe WatchSource:0}: Error finding container d260aab98481de9ce91264fba0b4bcff3062b11c51127d1582d69e54415c4cfe: Status 404 returned error can't find the container with id d260aab98481de9ce91264fba0b4bcff3062b11c51127d1582d69e54415c4cfe Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.803080 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b9gn8" Jan 29 12:05:38 crc kubenswrapper[4993]: W0129 12:05:38.821828 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd22ea793_b0f9_4b2c_8ea7_ed61728d900f.slice/crio-76d2426c78e1a1a00a91b2d380873756cdd1682bb80d9d3f949ee32405d2f83b WatchSource:0}: Error finding container 76d2426c78e1a1a00a91b2d380873756cdd1682bb80d9d3f949ee32405d2f83b: Status 404 returned error can't find the container with id 76d2426c78e1a1a00a91b2d380873756cdd1682bb80d9d3f949ee32405d2f83b Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.845500 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-f9pq9"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.846096 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.846793 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzr6r"] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.848471 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.848612 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.848662 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854529 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854570 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854570 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854650 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854577 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.854761 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.856501 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.858965 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859059 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859089 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859116 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t8gs\" (UniqueName: \"kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859138 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.859168 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:05:39.859139717 +0000 UTC m=+23.874269873 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859225 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sbzh\" (UniqueName: \"kubernetes.io/projected/0a56c564-0553-48b1-a375-c98f3901a0e2-kube-api-access-7sbzh\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859260 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859284 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-system-cni-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859364 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859404 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859452 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-os-release\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859471 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859491 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859519 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859585 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859618 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859710 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-binary-copy\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859771 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859797 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859830 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859858 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859966 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.859998 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860072 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.860167 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860244 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.860303 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:39.860285507 +0000 UTC m=+23.875415743 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860591 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860638 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860661 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860685 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-cnibin\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.860714 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.860841 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.860896 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:39.860880833 +0000 UTC m=+23.876010979 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.861658 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.870170 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.880480 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.890000 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.900456 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.909580 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.922937 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.932827 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.944311 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.953152 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961560 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961620 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961644 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961667 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961702 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961726 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961749 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961753 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961770 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961878 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961912 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-cnibin\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961955 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.961993 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962006 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962016 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962032 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962087 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:39.962070342 +0000 UTC m=+23.977200578 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962245 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961817 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962302 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-cnibin\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962306 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961882 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.961830 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962346 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962031 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t8gs\" (UniqueName: \"kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962392 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962404 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962427 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sbzh\" (UniqueName: \"kubernetes.io/projected/0a56c564-0553-48b1-a375-c98f3901a0e2-kube-api-access-7sbzh\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962460 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-system-cni-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962483 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962485 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962516 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962535 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962572 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962591 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-os-release\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962610 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962631 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962651 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962677 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962700 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962727 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-binary-copy\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962752 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962779 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962793 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-system-cni-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962799 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962824 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962918 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962947 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962959 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: E0129 12:05:38.962991 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:39.962980427 +0000 UTC m=+23.978110643 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.962999 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-os-release\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963020 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963031 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963028 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963087 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963116 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0a56c564-0553-48b1-a375-c98f3901a0e2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963135 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963730 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963736 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.963778 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.964019 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0a56c564-0553-48b1-a375-c98f3901a0e2-cni-binary-copy\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.968754 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.968923 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.979944 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.983374 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sbzh\" (UniqueName: \"kubernetes.io/projected/0a56c564-0553-48b1-a375-c98f3901a0e2-kube-api-access-7sbzh\") pod \"multus-additional-cni-plugins-f9pq9\" (UID: \"0a56c564-0553-48b1-a375-c98f3901a0e2\") " pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.984353 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t8gs\" (UniqueName: \"kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs\") pod \"ovnkube-node-pzr6r\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:38 crc kubenswrapper[4993]: I0129 12:05:38.990617 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.003769 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.014761 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.025550 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.037252 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.044429 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.069718 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.105941 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.145391 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.159926 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:19:17.241357183 +0000 UTC Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.188256 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.194593 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.195381 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.196587 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.197229 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.198290 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.198766 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.199377 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.200275 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.200862 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.201873 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.202414 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.203619 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.204237 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.204861 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.206036 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.206702 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.207802 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.208211 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.208769 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.210036 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.210592 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.211725 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.212306 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.213561 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.214065 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.214812 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.216294 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.216850 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.218015 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.218547 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.219528 4993 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.219652 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.221335 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.222536 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.223028 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.224728 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.225455 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.226487 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.227385 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.228522 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.229137 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.230263 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.230981 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.232137 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.232252 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.232889 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.233432 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.233991 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.234678 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.236205 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.236387 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.236769 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.237797 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.238385 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.239564 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.240523 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.241137 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.339060 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"225889fd3889840ab1c01c1dd9a0aa1ad8002f0e6a426e098d662f7517f0732a"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.341138 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.341635 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.343464 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" exitCode=255 Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.343539 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.343620 4993 scope.go:117] "RemoveContainer" containerID="15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.344918 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"33312ff181791b1819893a1d1cfd5060d78cff3cf0996930db8326b0096ec142"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.345798 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerStarted","Data":"2720ce404f897497db91effc28803faa2e8d1e63200a276690a7373305e03055"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.350205 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.350257 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.350268 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"d260aab98481de9ce91264fba0b4bcff3062b11c51127d1582d69e54415c4cfe"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.352995 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.353234 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d1088bb9f5ba6b194db2326573f6a89bc8a3ea76aed9474ffa35e18ea738b8d0"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.355280 4993 scope.go:117] "RemoveContainer" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.355421 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.355718 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.368610 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.389162 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerStarted","Data":"81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.389579 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerStarted","Data":"76d2426c78e1a1a00a91b2d380873756cdd1682bb80d9d3f949ee32405d2f83b"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.397533 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-clvrz" event={"ID":"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8","Type":"ContainerStarted","Data":"f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.397580 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-clvrz" event={"ID":"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8","Type":"ContainerStarted","Data":"5de047a797640c48c6c074ba58e3bed2a51d13d0c2b01d220e3d1fc3684a9c1c"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.400774 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.400810 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.400823 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4bfe996d6650c66ab8a29d99e7845fdd7a9d37296b0a7820b17ddcd5924f6efd"} Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.401400 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.419103 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.430766 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.446412 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.487755 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.526683 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.569616 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.610275 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.645616 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.688798 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.729401 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.774320 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.810521 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.849621 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.871702 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.871879 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.871914 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:05:41.871881646 +0000 UTC m=+25.887011782 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.871971 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.871993 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.872089 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:41.872066991 +0000 UTC m=+25.887197297 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.872209 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.872326 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:41.872298677 +0000 UTC m=+25.887428993 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.889412 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.930050 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.969059 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:39Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.972510 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:39 crc kubenswrapper[4993]: I0129 12:05:39.972580 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972729 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972759 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972775 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972803 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972831 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972835 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:41.972818019 +0000 UTC m=+25.987948145 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972844 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:39 crc kubenswrapper[4993]: E0129 12:05:39.972897 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:41.972880801 +0000 UTC m=+25.988010927 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.011500 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.082044 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.118518 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.135399 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.160212 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 15:24:55.206323718 +0000 UTC Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.168309 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.190220 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.190255 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.190294 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:40 crc kubenswrapper[4993]: E0129 12:05:40.190356 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:40 crc kubenswrapper[4993]: E0129 12:05:40.190428 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:40 crc kubenswrapper[4993]: E0129 12:05:40.190516 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.205930 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.250415 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://15b09ef6db468a5ac3aa1f21c350123702e9a39ae0d9078787aaa4be526a654a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:31Z\\\",\\\"message\\\":\\\"W0129 12:05:20.548024 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 12:05:20.548296 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769688320 cert, and key in /tmp/serving-cert-1503077017/serving-signer.crt, /tmp/serving-cert-1503077017/serving-signer.key\\\\nI0129 12:05:21.036706 1 observer_polling.go:159] Starting file observer\\\\nW0129 12:05:21.040895 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 12:05:21.041098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:21.043134 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1503077017/tls.crt::/tmp/serving-cert-1503077017/tls.key\\\\\\\"\\\\nF0129 12:05:31.671610 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.406228 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.408577 4993 scope.go:117] "RemoveContainer" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" Jan 29 12:05:40 crc kubenswrapper[4993]: E0129 12:05:40.408853 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.409062 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" exitCode=0 Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.409123 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.410664 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf" exitCode=0 Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.410710 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf"} Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.429948 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.449547 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.460296 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.479704 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.490993 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.504362 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.532992 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.568904 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.608994 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.646509 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.688217 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.727513 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.768160 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.811148 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.848027 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.886766 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.933213 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:40 crc kubenswrapper[4993]: I0129 12:05:40.970284 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.008087 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.049551 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.086479 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.128243 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.160818 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 08:07:29.000166353 +0000 UTC Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.173214 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.208112 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.248940 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.289814 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420825 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420894 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420907 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420918 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420927 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.420938 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.423560 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83" exitCode=0 Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.423677 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.425507 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee"} Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.446529 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.471823 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.483859 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.505620 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.522624 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.536580 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.574824 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.622494 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.653931 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.688319 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.730929 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.770981 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.815582 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.851830 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.889649 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.889743 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.889773 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.889862 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.889911 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:45.889898327 +0000 UTC m=+29.905028453 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.889940 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.890041 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:45.890013421 +0000 UTC m=+29.905143587 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.890150 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:05:45.890134074 +0000 UTC m=+29.905264240 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.891793 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.926675 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.970953 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:41Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.990567 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:41 crc kubenswrapper[4993]: I0129 12:05:41.990665 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990857 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990886 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990895 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990952 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990905 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.990999 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.991046 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:45.991024856 +0000 UTC m=+30.006155022 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:41 crc kubenswrapper[4993]: E0129 12:05:41.991156 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:45.991078967 +0000 UTC m=+30.006209133 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.012407 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.050835 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.092893 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.130830 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.161118 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:11:44.083675331 +0000 UTC Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.170828 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.189407 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.189481 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:42 crc kubenswrapper[4993]: E0129 12:05:42.189545 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.189611 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:42 crc kubenswrapper[4993]: E0129 12:05:42.189616 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:42 crc kubenswrapper[4993]: E0129 12:05:42.189689 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.209260 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.246335 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.295858 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.329044 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.430497 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14" exitCode=0 Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.430600 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14"} Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.444360 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.458393 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.473347 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.491475 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.526984 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.569622 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.606985 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.650881 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.688400 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.728123 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.767778 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.807395 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:42 crc kubenswrapper[4993]: I0129 12:05:42.849662 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:42Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.162273 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 04:14:23.143543204 +0000 UTC Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.437289 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.440002 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303" exitCode=0 Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.440031 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303"} Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.465696 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.478973 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.497033 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.521786 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.536689 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.549361 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.562331 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.573701 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.584360 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.598660 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.625726 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.640833 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.655025 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.701772 4993 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.705975 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.706027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.706040 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.706117 4993 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.712142 4993 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.712451 4993 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.713640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.713681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.713692 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.713709 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.713720 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.724438 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.727639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.727673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.727685 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.727700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.727711 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.897774 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.902172 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.902233 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.902243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.902263 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.902275 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.913807 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.916983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.917030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.917041 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.917062 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.917075 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.931197 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.934788 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.934820 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.934828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.934840 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.934849 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.947920 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:43Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:43 crc kubenswrapper[4993]: E0129 12:05:43.948127 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.949621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.949658 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.949669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.949684 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:43 crc kubenswrapper[4993]: I0129 12:05:43.949695 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:43Z","lastTransitionTime":"2026-01-29T12:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.052476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.052524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.052537 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.052556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.052570 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.155120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.155159 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.155167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.155195 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.155212 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.162735 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 18:58:36.726721652 +0000 UTC Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.190261 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.190273 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:44 crc kubenswrapper[4993]: E0129 12:05:44.190429 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.190283 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:44 crc kubenswrapper[4993]: E0129 12:05:44.190487 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:44 crc kubenswrapper[4993]: E0129 12:05:44.190532 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.257722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.257770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.257781 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.257797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.257808 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.360475 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.360513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.360524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.360539 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.360551 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.452602 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97" exitCode=0 Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.452698 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.462792 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.462841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.462854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.462875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.462888 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.474428 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.490418 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.503487 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.519060 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.531704 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.549240 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.565332 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.565384 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.565403 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.565431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.565450 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.573608 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.591289 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.606406 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.622693 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.635211 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.654360 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.668406 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.668453 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.668466 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.668485 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.668498 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.673146 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:44Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.771132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.771236 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.771249 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.771280 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.771294 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.874955 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.875401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.875415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.875436 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.875451 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.978284 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.978527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.978544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.978564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:44 crc kubenswrapper[4993]: I0129 12:05:44.978577 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:44Z","lastTransitionTime":"2026-01-29T12:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.081311 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.081768 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.081913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.082036 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.082140 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.163274 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 23:06:06.640431521 +0000 UTC Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.185375 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.185417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.185426 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.185443 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.185453 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.288173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.288248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.288262 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.288280 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.288292 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.390671 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.390717 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.390726 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.390741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.390751 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.461838 4993 generic.go:334] "Generic (PLEG): container finished" podID="0a56c564-0553-48b1-a375-c98f3901a0e2" containerID="3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014" exitCode=0 Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.461881 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerDied","Data":"3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.475273 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.487968 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.493156 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.493197 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.493206 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.493219 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.493229 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.501709 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.513515 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.527743 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.540166 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.553274 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.566370 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.577527 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.592354 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.597115 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.597168 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.597180 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.597239 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.597269 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.605938 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.622271 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.641030 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.700102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.700153 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.700165 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.700210 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.700232 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.803393 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.803440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.803454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.803475 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.803490 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.905893 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.905931 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.905939 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.905952 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.905962 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:45Z","lastTransitionTime":"2026-01-29T12:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.915468 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.915614 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:45 crc kubenswrapper[4993]: E0129 12:05:45.915639 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:05:53.915619487 +0000 UTC m=+37.930749613 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:45 crc kubenswrapper[4993]: I0129 12:05:45.915690 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:45 crc kubenswrapper[4993]: E0129 12:05:45.915693 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:45 crc kubenswrapper[4993]: E0129 12:05:45.915748 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:45 crc kubenswrapper[4993]: E0129 12:05:45.915752 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:53.91574325 +0000 UTC m=+37.930873366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:45 crc kubenswrapper[4993]: E0129 12:05:45.915808 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:53.915793221 +0000 UTC m=+37.930923347 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.008507 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.008548 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.008560 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.008576 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.008587 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.016357 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.016455 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016545 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016581 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016584 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016608 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016615 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016631 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016692 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:54.016672372 +0000 UTC m=+38.031802518 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.016715 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:54.016704683 +0000 UTC m=+38.031834819 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.111120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.111165 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.111178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.111226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.111258 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.164525 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 13:09:29.300931384 +0000 UTC Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.189890 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.189970 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.190062 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.190000 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.190178 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:46 crc kubenswrapper[4993]: E0129 12:05:46.190342 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.214790 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.214855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.214870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.214893 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.214909 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.317414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.317451 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.317461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.317477 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.317488 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.420295 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.420344 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.420359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.420379 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.420392 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.468560 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.468829 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.468865 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.472922 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" event={"ID":"0a56c564-0553-48b1-a375-c98f3901a0e2","Type":"ContainerStarted","Data":"d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.483010 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.492712 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.493078 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.494357 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.506621 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.522020 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.522071 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.522083 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.522103 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.522115 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.524964 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.537635 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.551353 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.566434 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.578556 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.590841 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.602472 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.614730 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.624136 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.624172 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.624202 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.624226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.624239 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.628255 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.637815 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.649011 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.663255 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.677817 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.690492 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.700158 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.721815 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.726916 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.726970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.726987 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.727011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.727027 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.742645 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.755372 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.771102 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.785700 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.800752 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.818862 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.829591 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.829696 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.829710 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.829727 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.829740 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.834118 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:46Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.931993 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.932029 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.932038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.932054 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.932067 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:46Z","lastTransitionTime":"2026-01-29T12:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:46 crc kubenswrapper[4993]: I0129 12:05:46.961346 4993 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.035461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.035515 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.035529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.035550 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.035564 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.138970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.139027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.139039 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.139057 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.139072 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.165668 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 05:37:15.699782109 +0000 UTC Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.208576 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.233699 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.242321 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.242405 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.242428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.242464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.242490 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.251984 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.271223 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.285054 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.300842 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.313353 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.328740 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.347008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.347053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.347064 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.347077 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.347088 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.350823 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.365406 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.379772 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.393465 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.404674 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.450148 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.450203 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.450216 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.450236 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.450247 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.476247 4993 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.552898 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.552973 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.552996 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.553027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.553055 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.656157 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.656228 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.656243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.656263 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.656278 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.759481 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.759546 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.759563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.759589 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.759606 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.861949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.862001 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.862018 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.862043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.862060 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.964492 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.964548 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.964561 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.964581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:47 crc kubenswrapper[4993]: I0129 12:05:47.964599 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:47Z","lastTransitionTime":"2026-01-29T12:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.066700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.066744 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.066754 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.066769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.066778 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.166455 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:04:28.966961914 +0000 UTC Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.168947 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.168970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.168978 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.168991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.169000 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.189751 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.189786 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.189825 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:48 crc kubenswrapper[4993]: E0129 12:05:48.189910 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:48 crc kubenswrapper[4993]: E0129 12:05:48.190031 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:48 crc kubenswrapper[4993]: E0129 12:05:48.190097 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.271268 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.271316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.271330 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.271347 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.271358 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.374557 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.374599 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.374609 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.374623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.374636 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.476905 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.476949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.476959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.476974 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.476986 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.478594 4993 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.555651 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.556491 4993 scope.go:117] "RemoveContainer" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" Jan 29 12:05:48 crc kubenswrapper[4993]: E0129 12:05:48.556676 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.579041 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.579070 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.579081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.579096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.579106 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.681456 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.681495 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.681506 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.681521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.681531 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.783552 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.783851 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.784055 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.784264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.784404 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.886689 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.886732 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.886743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.886760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.886771 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.989724 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.989767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.989777 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.989793 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:48 crc kubenswrapper[4993]: I0129 12:05:48.989804 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:48Z","lastTransitionTime":"2026-01-29T12:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.092313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.092361 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.092371 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.092385 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.092395 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.166618 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 04:31:35.745604551 +0000 UTC Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.193741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.193774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.193783 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.193795 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.193805 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.296862 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.296907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.296924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.296942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.296957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.400431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.400504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.400531 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.400562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.400581 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.483885 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/0.log" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.488460 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8" exitCode=1 Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.488508 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.489310 4993 scope.go:117] "RemoveContainer" containerID="6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.503717 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.503785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.503805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.503830 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.503851 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.510712 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.526587 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.546476 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.570816 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.589962 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.607007 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.607107 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.607133 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.607624 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.607948 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.614161 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.630704 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.645765 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.663647 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.677504 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.691872 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.708844 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.710804 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.710841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.710850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.710864 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.710874 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.724860 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:49Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.814242 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.814316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.814354 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.814386 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.814406 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.916644 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.916705 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.916723 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.916747 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:49 crc kubenswrapper[4993]: I0129 12:05:49.916767 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:49Z","lastTransitionTime":"2026-01-29T12:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.019254 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.019291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.019300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.019314 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.019326 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.121926 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.121965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.121978 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.121994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.122003 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.167295 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 02:43:01.198712438 +0000 UTC Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.189611 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.189680 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:50 crc kubenswrapper[4993]: E0129 12:05:50.189732 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.189685 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:50 crc kubenswrapper[4993]: E0129 12:05:50.189921 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:50 crc kubenswrapper[4993]: E0129 12:05:50.190104 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.224274 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.224319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.224331 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.224350 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.224362 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.326739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.326781 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.326793 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.326807 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.326817 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.429428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.429469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.429481 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.429498 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.429510 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.494795 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/0.log" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.498247 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.498396 4993 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.513586 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.528116 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.532168 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.532229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.532240 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.532258 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.532273 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.547525 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.561871 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.575905 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.590534 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.602986 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.620336 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.634736 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.634769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.634776 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.634788 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.634797 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.640413 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.657940 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.672101 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.684482 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.700163 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.736991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.737028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.737037 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.737052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.737061 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.839296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.839329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.839337 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.839351 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.839360 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.942650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.942718 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.942736 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.942760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:50 crc kubenswrapper[4993]: I0129 12:05:50.942781 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:50Z","lastTransitionTime":"2026-01-29T12:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.044677 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.044730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.044741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.044759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.044771 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.147729 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.147784 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.147794 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.147816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.147829 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.167453 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 13:39:45.200024034 +0000 UTC Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.251149 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.251229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.251241 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.251258 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.251269 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.353447 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.353486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.353494 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.353507 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.353517 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.455940 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.455987 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.455998 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.456015 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.456028 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.503347 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/1.log" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.503844 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/0.log" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.506928 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1" exitCode=1 Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.506959 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.506998 4993 scope.go:117] "RemoveContainer" containerID="6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.507586 4993 scope.go:117] "RemoveContainer" containerID="0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1" Jan 29 12:05:51 crc kubenswrapper[4993]: E0129 12:05:51.507728 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.526435 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.546645 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.558839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.558882 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.558892 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.558913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.558927 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.564948 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.585341 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.596968 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.632968 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.652235 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.661687 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.661736 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.661750 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.661772 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.661789 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.667714 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.686622 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.699892 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.704542 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd"] Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.704977 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.706678 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.707003 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.719021 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.730949 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.748940 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764698 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764779 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764784 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764821 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.764932 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.776030 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-env-overrides\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.776072 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.776223 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61f75260-59f9-40c7-868c-4802106efa65-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.776255 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shcp2\" (UniqueName: \"kubernetes.io/projected/61f75260-59f9-40c7-868c-4802106efa65-kube-api-access-shcp2\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.776809 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:51 crc kubenswrapper[4993]: I0129 12:05:51.789320 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.800782 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.811128 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.825562 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.842828 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.853830 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.867224 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.867265 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.867281 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.867300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.867316 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.868774 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.877609 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61f75260-59f9-40c7-868c-4802106efa65-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.877652 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shcp2\" (UniqueName: \"kubernetes.io/projected/61f75260-59f9-40c7-868c-4802106efa65-kube-api-access-shcp2\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.877713 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-env-overrides\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.877747 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.878380 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-env-overrides\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.878686 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/61f75260-59f9-40c7-868c-4802106efa65-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.882535 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.883769 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/61f75260-59f9-40c7-868c-4802106efa65-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.894409 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shcp2\" (UniqueName: \"kubernetes.io/projected/61f75260-59f9-40c7-868c-4802106efa65-kube-api-access-shcp2\") pod \"ovnkube-control-plane-749d76644c-w74pd\" (UID: \"61f75260-59f9-40c7-868c-4802106efa65\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.897892 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.910999 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.922255 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.933538 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:51Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.969206 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.969232 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.969243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.969278 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:51.969290 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:51Z","lastTransitionTime":"2026-01-29T12:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.019084 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.167923 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 19:20:57.90790886 +0000 UTC Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.189413 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.189441 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:52 crc kubenswrapper[4993]: E0129 12:05:52.189541 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.189573 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:52 crc kubenswrapper[4993]: E0129 12:05:52.189810 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:52 crc kubenswrapper[4993]: E0129 12:05:52.189974 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.422475 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-qn5kr"] Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.422896 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.426310 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.426691 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.426931 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.428700 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.994342 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.994447 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.994516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.994580 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:52 crc kubenswrapper[4993]: I0129 12:05:52.994634 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:52Z","lastTransitionTime":"2026-01-29T12:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.000042 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" event={"ID":"61f75260-59f9-40c7-868c-4802106efa65","Type":"ContainerStarted","Data":"0dd0d1f6221bb30f6d2899aaca04238620c7378bbd31b438966a025bb16f565d"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.001156 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.002893 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/1.log" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.015873 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.028455 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.040656 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.053888 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.069308 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.087663 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.089989 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8bf1534-f3e0-4302-a00d-17c0721fd168-host\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.090071 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ghg4\" (UniqueName: \"kubernetes.io/projected/d8bf1534-f3e0-4302-a00d-17c0721fd168-kube-api-access-7ghg4\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.090123 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d8bf1534-f3e0-4302-a00d-17c0721fd168-serviceca\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.097615 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.097648 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.097660 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.097676 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.097687 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.106425 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.120589 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.131146 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.145684 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.157931 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.168199 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 10:53:58.899534995 +0000 UTC Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.172135 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.190979 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d8bf1534-f3e0-4302-a00d-17c0721fd168-serviceca\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.191009 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8bf1534-f3e0-4302-a00d-17c0721fd168-host\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.191056 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ghg4\" (UniqueName: \"kubernetes.io/projected/d8bf1534-f3e0-4302-a00d-17c0721fd168-kube-api-access-7ghg4\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.191201 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8bf1534-f3e0-4302-a00d-17c0721fd168-host\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.191381 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.192415 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d8bf1534-f3e0-4302-a00d-17c0721fd168-serviceca\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.199888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.200143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.200317 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.200422 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.200484 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.203872 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.215447 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ghg4\" (UniqueName: \"kubernetes.io/projected/d8bf1534-f3e0-4302-a00d-17c0721fd168-kube-api-access-7ghg4\") pod \"node-ca-qn5kr\" (UID: \"d8bf1534-f3e0-4302-a00d-17c0721fd168\") " pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.302884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.302934 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.302943 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.302958 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.302967 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.335388 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vdgbx"] Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.335979 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: E0129 12:05:53.336067 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.352175 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qn5kr" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.354120 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.370866 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: W0129 12:05:53.375898 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8bf1534_f3e0_4302_a00d_17c0721fd168.slice/crio-3310c3adfb9b6865c146ba9024ed7d98a0280ba333f6fb91b10b6c5ca3acfbda WatchSource:0}: Error finding container 3310c3adfb9b6865c146ba9024ed7d98a0280ba333f6fb91b10b6c5ca3acfbda: Status 404 returned error can't find the container with id 3310c3adfb9b6865c146ba9024ed7d98a0280ba333f6fb91b10b6c5ca3acfbda Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.393773 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.394111 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dsr2\" (UniqueName: \"kubernetes.io/projected/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-kube-api-access-7dsr2\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.404036 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.406110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.406147 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.406156 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.406172 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.406193 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.420029 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.432846 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.448072 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.460081 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.471212 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.481918 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.491719 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.494753 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dsr2\" (UniqueName: \"kubernetes.io/projected/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-kube-api-access-7dsr2\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.494815 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: E0129 12:05:53.494922 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:53 crc kubenswrapper[4993]: E0129 12:05:53.494976 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:53.994959964 +0000 UTC m=+38.010090090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.505729 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.508778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.508812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.508826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.508845 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.508857 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.511376 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dsr2\" (UniqueName: \"kubernetes.io/projected/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-kube-api-access-7dsr2\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.522035 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.533496 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.541289 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.552275 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.563974 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:53Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.610662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.610698 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.610707 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.610721 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.610732 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.712964 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.713020 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.713037 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.713060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.713078 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.816713 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.816797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.816824 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.816929 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.817001 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.920618 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.921045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.921086 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.921111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.921127 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:53Z","lastTransitionTime":"2026-01-29T12:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:53 crc kubenswrapper[4993]: I0129 12:05:53.998807 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:53.999037 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999077 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:06:09.999039053 +0000 UTC m=+54.014169219 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999172 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:53.999182 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999299 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:09.999273989 +0000 UTC m=+54.014404165 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999342 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999459 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:09.999438893 +0000 UTC m=+54.014569049 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:53.999341 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999487 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:53.999559 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:54.999537866 +0000 UTC m=+39.014668032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.012579 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" event={"ID":"61f75260-59f9-40c7-868c-4802106efa65","Type":"ContainerStarted","Data":"f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.012639 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" event={"ID":"61f75260-59f9-40c7-868c-4802106efa65","Type":"ContainerStarted","Data":"358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.014733 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qn5kr" event={"ID":"d8bf1534-f3e0-4302-a00d-17c0721fd168","Type":"ContainerStarted","Data":"328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.014784 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qn5kr" event={"ID":"d8bf1534-f3e0-4302-a00d-17c0721fd168","Type":"ContainerStarted","Data":"3310c3adfb9b6865c146ba9024ed7d98a0280ba333f6fb91b10b6c5ca3acfbda"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.023441 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.023502 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.023525 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.023554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.023576 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.027979 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.042017 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.054480 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.068616 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.084218 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.100852 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.101003 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.101029 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.101045 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.101103 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:10.101083465 +0000 UTC m=+54.116213691 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.101930 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.102074 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.102109 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.102129 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.102221 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:10.102163954 +0000 UTC m=+54.117294120 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.108756 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.127286 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.127357 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.127381 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.127412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.127436 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.136053 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.148810 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.162298 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.168941 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 11:12:56.984397077 +0000 UTC Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.181024 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.185504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.185578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.185602 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.185631 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.185656 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.189580 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.189653 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.189580 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.189772 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.189914 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.190021 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.199934 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.199823 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.205040 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.205093 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.205111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.205134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.205151 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.215593 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.224582 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.228772 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.228813 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.228828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.228847 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.228863 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.229581 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.240675 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.243205 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.246662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.246727 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.246743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.246763 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.246778 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.256411 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.261711 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.264875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.264928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.264944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.264973 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.264989 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.271716 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.282397 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: E0129 12:05:54.282533 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.284080 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.284108 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.284119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.284137 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.284148 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.287655 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.306374 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.322046 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.336405 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.348287 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.365421 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.379012 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.386985 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.387053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.387067 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.387085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.387098 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.398774 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.418133 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6505d6863f7e5cce3abad1020d254105fe0fa123b45672dbb6d53c15eb142ac8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:48Z\\\",\\\"message\\\":\\\"ent-go/informers/factory.go:160\\\\nI0129 12:05:48.547526 6244 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0129 12:05:48.547680 6244 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0129 12:05:48.548430 6244 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 12:05:48.548472 6244 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0129 12:05:48.548479 6244 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0129 12:05:48.548497 6244 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0129 12:05:48.548510 6244 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0129 12:05:48.548514 6244 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0129 12:05:48.548541 6244 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0129 12:05:48.548553 6244 factory.go:656] Stopping watch factory\\\\nI0129 12:05:48.548540 6244 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0129 12:05:48.548568 6244 ovnkube.go:599] Stopped ovnkube\\\\nI0129 12:05:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.428803 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.440512 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.452944 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.464826 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.475399 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.487047 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.489448 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.489484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.489497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.489515 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.489528 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.498697 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:54Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.592737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.592767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.592775 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.592805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.592815 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.695981 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.696050 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.696062 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.696081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.696094 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.798737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.798806 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.798845 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.798880 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.798907 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.901058 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.901136 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.901160 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.901226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:54 crc kubenswrapper[4993]: I0129 12:05:54.901251 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:54Z","lastTransitionTime":"2026-01-29T12:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.004869 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.005036 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.005068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.005101 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.005125 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.011938 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:55 crc kubenswrapper[4993]: E0129 12:05:55.012071 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:55 crc kubenswrapper[4993]: E0129 12:05:55.012144 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:05:57.012122312 +0000 UTC m=+41.027252448 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.107984 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.108098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.108117 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.108143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.108161 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.169315 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 20:19:55.957802988 +0000 UTC Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.189839 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:55 crc kubenswrapper[4993]: E0129 12:05:55.190029 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.210983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.211049 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.211066 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.211089 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.211107 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.314745 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.314827 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.314849 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.314878 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.314910 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.417431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.417492 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.417513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.417538 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.417557 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.519608 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.519677 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.519689 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.519706 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.519718 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.622519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.622595 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.622616 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.622653 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.622676 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.724913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.724963 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.724979 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.724997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.725010 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.828081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.828143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.828154 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.828171 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.828203 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.926817 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.927810 4993 scope.go:117] "RemoveContainer" containerID="0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1" Jan 29 12:05:55 crc kubenswrapper[4993]: E0129 12:05:55.928025 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.930652 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.930701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.930713 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.930731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.930743 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:55Z","lastTransitionTime":"2026-01-29T12:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.944832 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.959890 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:55 crc kubenswrapper[4993]: I0129 12:05:55.976901 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.003018 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.018328 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.032902 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.032942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.032950 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.032964 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.032973 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.035668 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.050061 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.062362 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.078612 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.090475 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.099946 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.112771 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.123801 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.134930 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.134969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.134977 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.134992 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.135004 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.135637 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.148674 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.163865 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:56Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.170216 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 19:36:37.072478064 +0000 UTC Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.189845 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.189903 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.189885 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:56 crc kubenswrapper[4993]: E0129 12:05:56.190031 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:56 crc kubenswrapper[4993]: E0129 12:05:56.190166 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:56 crc kubenswrapper[4993]: E0129 12:05:56.190317 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.237820 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.237863 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.237873 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.237888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.237901 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.340372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.340427 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.340439 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.340458 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.340470 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.443911 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.443980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.443999 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.444024 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.444042 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.547327 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.547409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.547447 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.547491 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.547516 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.650214 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.650258 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.650273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.650295 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.650310 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.752214 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.752255 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.752272 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.752289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.752301 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.854667 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.854706 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.854716 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.854731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.854742 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.957480 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.957537 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.957545 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.957578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:56 crc kubenswrapper[4993]: I0129 12:05:56.957590 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:56Z","lastTransitionTime":"2026-01-29T12:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.032274 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:57 crc kubenswrapper[4993]: E0129 12:05:57.032467 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:57 crc kubenswrapper[4993]: E0129 12:05:57.032540 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:01.032522637 +0000 UTC m=+45.047652763 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.060052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.060093 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.060109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.060135 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.060152 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.163010 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.163046 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.163056 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.163067 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.163077 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.171206 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 17:53:23.35247353 +0000 UTC Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.189572 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:57 crc kubenswrapper[4993]: E0129 12:05:57.189706 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.212359 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.231493 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.247049 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.261779 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.265445 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.265501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.265519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.265544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.265561 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.273320 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.289439 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.301466 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.314531 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.327864 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.338414 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.350125 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.359734 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.367721 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.367769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.367783 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.367801 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.367812 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.372661 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.392443 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.404292 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.414159 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.471078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.471244 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.471269 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.471288 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.471301 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.574644 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.574709 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.574727 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.574757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.574782 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.677764 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.677796 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.677805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.677818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.677827 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.781242 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.781289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.781305 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.781329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.781345 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.884778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.884839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.884858 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.884913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.884934 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.988128 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.988178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.988209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.988226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:57 crc kubenswrapper[4993]: I0129 12:05:57.988238 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:57Z","lastTransitionTime":"2026-01-29T12:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.090367 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.090424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.090435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.090455 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.090464 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.171921 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:14:51.485659 +0000 UTC Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.190395 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.190445 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.190406 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:05:58 crc kubenswrapper[4993]: E0129 12:05:58.190559 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:05:58 crc kubenswrapper[4993]: E0129 12:05:58.190724 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:05:58 crc kubenswrapper[4993]: E0129 12:05:58.190782 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.192568 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.192604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.192617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.192632 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.192643 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.295688 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.295728 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.295741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.295757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.295770 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.399043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.399091 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.399102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.399121 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.399133 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.502128 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.502386 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.502397 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.502408 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.502416 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.604784 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.604866 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.604891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.604922 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.604943 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.707860 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.707925 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.707951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.707983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.708006 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.811085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.811148 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.811169 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.811467 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.811528 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.914811 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.914865 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.914884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.914911 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:58 crc kubenswrapper[4993]: I0129 12:05:58.914929 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:58Z","lastTransitionTime":"2026-01-29T12:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.017431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.017468 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.017477 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.017490 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.017499 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.120142 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.120178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.120208 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.120223 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.120232 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.172539 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 06:05:47.589841337 +0000 UTC Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.189897 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:05:59 crc kubenswrapper[4993]: E0129 12:05:59.190023 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.222841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.222874 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.222884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.222897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.222906 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.325336 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.325373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.325386 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.325403 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.325423 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.427684 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.427721 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.427730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.427744 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.427756 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.530454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.530488 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.530497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.530511 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.530519 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.632713 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.632748 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.632757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.632778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.632787 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.735566 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.735612 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.735626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.735646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.735658 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.838078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.838147 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.838162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.838211 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.838228 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.940633 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.940683 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.940693 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.940707 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:05:59 crc kubenswrapper[4993]: I0129 12:05:59.940717 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:05:59Z","lastTransitionTime":"2026-01-29T12:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.043225 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.043273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.043282 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.043296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.043308 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.145933 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.145994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.146227 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.146274 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.146296 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.173504 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 00:20:09.168184765 +0000 UTC Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.189884 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.189961 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.190016 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:00 crc kubenswrapper[4993]: E0129 12:06:00.190176 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:00 crc kubenswrapper[4993]: E0129 12:06:00.190402 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:00 crc kubenswrapper[4993]: E0129 12:06:00.190575 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.249561 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.249624 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.249638 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.249655 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.249668 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.352774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.352859 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.352869 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.352883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.352893 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.455874 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.455938 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.455960 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.455991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.456014 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.559473 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.559557 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.559582 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.559612 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.559633 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.663103 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.663143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.663157 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.663177 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.663210 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.766177 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.766579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.766778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.766956 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.767137 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.870606 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.870691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.870716 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.870747 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.870772 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.973867 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.973917 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.973931 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.973948 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:00 crc kubenswrapper[4993]: I0129 12:06:00.973959 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:00Z","lastTransitionTime":"2026-01-29T12:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.074929 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:01 crc kubenswrapper[4993]: E0129 12:06:01.075149 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:01 crc kubenswrapper[4993]: E0129 12:06:01.075322 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:09.075291708 +0000 UTC m=+53.090421844 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.076147 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.076251 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.076273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.076302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.076322 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.173801 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 06:01:03.077634779 +0000 UTC Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.178151 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.178246 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.178266 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.178289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.178305 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.190578 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:01 crc kubenswrapper[4993]: E0129 12:06:01.190689 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.281234 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.281264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.281272 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.281286 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.281295 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.388096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.388157 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.388170 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.388213 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.388226 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.490613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.490663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.490677 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.490699 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.490713 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.593102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.593141 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.593151 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.593167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.593212 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.695317 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.695623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.695739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.695891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.696026 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.799409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.799489 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.799513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.799543 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.799567 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.901554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.901599 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.901607 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.901624 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:01 crc kubenswrapper[4993]: I0129 12:06:01.901637 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:01Z","lastTransitionTime":"2026-01-29T12:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.003932 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.003995 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.004019 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.004052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.004077 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.107501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.107575 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.107613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.107647 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.107671 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.174893 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 14:40:17.935921875 +0000 UTC Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.189425 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.189550 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:02 crc kubenswrapper[4993]: E0129 12:06:02.189548 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.189603 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:02 crc kubenswrapper[4993]: E0129 12:06:02.189666 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:02 crc kubenswrapper[4993]: E0129 12:06:02.189886 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.190287 4993 scope.go:117] "RemoveContainer" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.210424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.210466 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.210476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.210497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.210511 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.312119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.312162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.312180 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.312216 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.312235 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.415291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.415349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.415361 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.415378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.415390 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.517205 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.517248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.517257 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.517271 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.517281 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.620319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.620397 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.620410 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.620432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.620447 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.722423 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.722453 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.722461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.722473 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.722516 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.825255 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.825698 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.826045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.826160 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.826281 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.928803 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.928857 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.928870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.928891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:02 crc kubenswrapper[4993]: I0129 12:06:02.928904 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:02Z","lastTransitionTime":"2026-01-29T12:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.031756 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.031799 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.031810 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.031826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.031836 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.047066 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.050240 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.050719 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.064509 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.078986 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.090718 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.105981 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.124654 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.134086 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.134118 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.134127 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.134142 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.134152 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.138105 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.153687 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.164984 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.175570 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 06:54:26.779432421 +0000 UTC Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.176982 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.189074 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.190243 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:03 crc kubenswrapper[4993]: E0129 12:06:03.190338 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.198301 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.208805 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.219853 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.230891 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.236347 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.236392 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.236403 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.236449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.236459 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.241739 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.252802 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:03Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.339124 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.339203 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.339215 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.339237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.339253 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.441891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.441927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.441935 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.441951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.441960 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.544731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.544785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.544803 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.544826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.544842 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.647288 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.647365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.647399 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.647432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.647454 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.750296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.750342 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.750353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.750366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.750377 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.853790 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.853855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.853872 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.853897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.853913 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.956487 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.956537 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.956576 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.956599 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:03 crc kubenswrapper[4993]: I0129 12:06:03.956617 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:03Z","lastTransitionTime":"2026-01-29T12:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.059532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.059595 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.059613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.059634 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.059648 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.163198 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.163235 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.163243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.163257 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.163265 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.176261 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 06:39:34.656238703 +0000 UTC Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.189946 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.189942 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.190104 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.190127 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.189978 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.190178 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.266687 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.266760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.266777 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.266800 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.266818 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.367247 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.367356 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.367381 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.367414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.367439 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.388311 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:04Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.392564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.392613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.392631 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.392650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.392664 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.410903 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:04Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.416949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.417020 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.417039 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.417066 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.417096 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.429833 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:04Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.434556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.434589 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.434598 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.434612 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.434622 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.448476 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:04Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.451876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.451953 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.451972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.451989 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.452002 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.465125 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:04Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:04 crc kubenswrapper[4993]: E0129 12:06:04.465295 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.466723 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.466753 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.466764 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.466779 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.466791 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.569813 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.569898 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.569920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.569951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.569970 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.673245 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.673314 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.673335 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.673363 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.673381 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.776802 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.776842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.776853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.776869 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.776880 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.879674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.879731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.879744 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.879761 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.879773 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.983473 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.983534 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.983553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.983579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:04 crc kubenswrapper[4993]: I0129 12:06:04.983597 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:04Z","lastTransitionTime":"2026-01-29T12:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.086493 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.086527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.086540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.086553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.086562 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.176520 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 08:10:52.054733648 +0000 UTC Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.189463 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:05 crc kubenswrapper[4993]: E0129 12:06:05.189737 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.189980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.190032 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.190048 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.190066 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.190080 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.292967 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.293054 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.293078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.293112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.293135 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.395833 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.395928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.395948 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.395970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.395986 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.498670 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.498724 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.498737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.498754 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.498769 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.601725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.601789 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.601801 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.601818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.601830 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.704509 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.704563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.704575 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.704597 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.704609 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.807021 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.807105 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.807126 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.807152 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.807170 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.910293 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.910361 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.910383 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.910412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:05 crc kubenswrapper[4993]: I0129 12:06:05.910434 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:05Z","lastTransitionTime":"2026-01-29T12:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.013777 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.013828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.013839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.013899 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.013923 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.116434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.116474 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.116482 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.116496 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.116505 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.176770 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 15:55:18.99266623 +0000 UTC Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.190039 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.190131 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:06 crc kubenswrapper[4993]: E0129 12:06:06.190152 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.190220 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:06 crc kubenswrapper[4993]: E0129 12:06:06.190291 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:06 crc kubenswrapper[4993]: E0129 12:06:06.190334 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.219276 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.219351 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.219364 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.219378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.219387 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.322416 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.322489 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.322512 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.322541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.322563 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.425674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.425725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.425737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.425755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.425770 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.528779 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.528850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.528864 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.528881 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.528926 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.631558 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.631812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.631936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.632029 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.632123 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.735470 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.735507 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.735519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.735534 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.735544 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.837974 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.838022 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.838038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.838056 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.838068 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.941535 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.941593 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.941610 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.941636 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.941652 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:06Z","lastTransitionTime":"2026-01-29T12:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.946966 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.960118 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.967306 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:06Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:06 crc kubenswrapper[4993]: I0129 12:06:06.987349 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:06Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.006465 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.022261 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.044606 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.044257 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.045054 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.045068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.045085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.045096 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.065127 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.077757 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.091958 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.106153 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.122023 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.136000 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.147122 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.147159 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.147167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.147199 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.147209 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.153838 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.164882 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.176978 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 16:34:42.43397626 +0000 UTC Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.182768 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.189927 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:07 crc kubenswrapper[4993]: E0129 12:06:07.190164 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.197761 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.210042 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.225236 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.240019 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.250537 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.250876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.251035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.251213 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.251354 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.251233 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.266449 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.283056 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.298036 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.317013 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.329881 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.344042 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.354081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.354154 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.354167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.354220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.354237 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.360400 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.376665 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.391284 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.405320 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.415529 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.426103 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.436899 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.449065 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.456997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.457029 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.457053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.457066 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.457074 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.558769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.558818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.558828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.558841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.558849 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.661159 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.661226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.661235 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.661247 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.661258 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.764442 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.764483 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.764495 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.764511 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.764523 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.866580 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.866675 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.866690 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.866707 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.866721 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.969660 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.969702 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.969711 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.969725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:07 crc kubenswrapper[4993]: I0129 12:06:07.969734 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:07Z","lastTransitionTime":"2026-01-29T12:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.073009 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.073088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.073105 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.073127 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.073163 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.176746 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.176842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.176861 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.176884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.176934 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.177216 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 03:59:35.365160674 +0000 UTC Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.190101 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.190097 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:08 crc kubenswrapper[4993]: E0129 12:06:08.190294 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.190108 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:08 crc kubenswrapper[4993]: E0129 12:06:08.190403 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:08 crc kubenswrapper[4993]: E0129 12:06:08.190496 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.279379 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.279423 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.279431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.279449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.279458 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.381850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.381921 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.381942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.381965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.381982 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.485179 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.485297 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.485331 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.485372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.485395 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.588365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.588420 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.588437 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.588462 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.588481 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.691533 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.691580 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.691589 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.691604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.691613 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.794350 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.794419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.794436 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.794464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.794506 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.897239 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.897296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.897313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.897334 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:08 crc kubenswrapper[4993]: I0129 12:06:08.897353 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:08Z","lastTransitionTime":"2026-01-29T12:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.000292 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.000367 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.000389 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.000417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.000443 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.102695 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.102734 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.102743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.102759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.102769 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.163844 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:09 crc kubenswrapper[4993]: E0129 12:06:09.164011 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:09 crc kubenswrapper[4993]: E0129 12:06:09.164093 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:25.164074767 +0000 UTC m=+69.179204893 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.178017 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 04:39:08.85132574 +0000 UTC Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.190519 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:09 crc kubenswrapper[4993]: E0129 12:06:09.190694 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.204907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.204954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.204966 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.204980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.204992 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.307617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.307671 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.307683 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.307706 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.307718 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.410371 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.410418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.410430 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.410446 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.410458 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.513158 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.513241 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.513253 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.513270 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.513281 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.617106 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.617458 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.617469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.617484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.617493 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.720291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.720365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.720384 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.720406 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.720422 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.822992 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.823034 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.823045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.823060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.823070 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.926028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.926085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.926100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.926120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:09 crc kubenswrapper[4993]: I0129 12:06:09.926136 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:09Z","lastTransitionTime":"2026-01-29T12:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.028515 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.028593 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.028661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.028696 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.028718 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.072589 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.072757 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:06:42.07272507 +0000 UTC m=+86.087855236 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.072933 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.073014 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.073107 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.073171 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.073233 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:42.073174062 +0000 UTC m=+86.088304228 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.073266 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:42.073251054 +0000 UTC m=+86.088381220 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.132075 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.132124 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.132135 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.132149 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.132159 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.173796 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.173874 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174044 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174075 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174081 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174151 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174177 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174094 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174329 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:42.174287971 +0000 UTC m=+86.189418137 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.174369 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:42.174351342 +0000 UTC m=+86.189481618 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.178858 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 17:13:56.514978579 +0000 UTC Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.190228 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.190281 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.190244 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.190413 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.190520 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:10 crc kubenswrapper[4993]: E0129 12:06:10.190670 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.234879 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.234947 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.234969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.234999 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.235020 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.337706 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.337751 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.337760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.337775 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.337784 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.441120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.441234 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.441258 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.441281 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.441516 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.543936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.544008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.544035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.544097 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.544122 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.646618 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.646653 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.646664 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.646680 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.646690 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.748899 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.748934 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.748941 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.748954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.748963 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.852525 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.852615 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.852666 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.852691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.852746 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.956717 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.956789 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.956812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.956841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:10 crc kubenswrapper[4993]: I0129 12:06:10.956864 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:10Z","lastTransitionTime":"2026-01-29T12:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.059863 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.059928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.059946 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.059969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.059988 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.163254 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.163346 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.163373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.163404 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.163437 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.179053 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:06:03.834216378 +0000 UTC Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.189765 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:11 crc kubenswrapper[4993]: E0129 12:06:11.190395 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.191035 4993 scope.go:117] "RemoveContainer" containerID="0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.266132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.266223 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.266239 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.266263 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.266280 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.368904 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.368936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.368944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.368959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.368968 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.471289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.471332 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.471343 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.471362 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.471375 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.573613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.573651 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.573663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.573679 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.573690 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.676541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.676591 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.676603 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.676621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.676633 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.778310 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.778344 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.778353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.778366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.778374 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.880264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.880305 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.880316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.880334 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.880346 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.983114 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.983227 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.983255 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.983285 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:11 crc kubenswrapper[4993]: I0129 12:06:11.983309 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:11Z","lastTransitionTime":"2026-01-29T12:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.079476 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/2.log" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.080579 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/1.log" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.083996 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9" exitCode=1 Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.084041 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.084120 4993 scope.go:117] "RemoveContainer" containerID="0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085314 4993 scope.go:117] "RemoveContainer" containerID="e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9" Jan 29 12:06:12 crc kubenswrapper[4993]: E0129 12:06:12.085762 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085848 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085898 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085921 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.085977 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.105263 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.122863 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.138030 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.156171 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.180068 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 22:27:34.163359811 +0000 UTC Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.181735 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.188777 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.188838 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.188854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.188881 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.188917 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.189429 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.189473 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.189504 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:12 crc kubenswrapper[4993]: E0129 12:06:12.189550 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:12 crc kubenswrapper[4993]: E0129 12:06:12.189688 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:12 crc kubenswrapper[4993]: E0129 12:06:12.189775 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.196778 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.213323 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.225694 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.239779 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.258927 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.270540 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.284574 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.291862 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.291892 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.291900 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.291914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.291923 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.301408 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.313572 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.324675 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.337000 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.346531 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:12Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.394428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.394464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.394472 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.394487 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.394496 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.496514 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.496549 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.496560 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.496577 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.496588 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.599248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.599291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.599304 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.599322 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.599334 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.702366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.702641 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.702730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.702810 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.702934 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.805164 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.805245 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.805262 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.805285 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.805306 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.908434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.908499 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.908518 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.908543 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:12 crc kubenswrapper[4993]: I0129 12:06:12.908561 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:12Z","lastTransitionTime":"2026-01-29T12:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.011552 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.011638 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.011661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.011691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.011710 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.091697 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/2.log" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.113918 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.113947 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.113956 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.113968 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.113976 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.181007 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 03:20:00.538315644 +0000 UTC Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.190678 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:13 crc kubenswrapper[4993]: E0129 12:06:13.190964 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.216530 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.216562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.216572 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.216587 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.216599 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.319419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.319547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.319650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.319679 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.319694 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.422446 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.422480 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.422491 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.422507 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.422520 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.525471 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.525510 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.525521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.525538 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.525549 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.629083 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.629217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.629259 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.629290 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.629311 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.732253 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.732368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.732390 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.732418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.732440 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.835467 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.835529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.835548 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.835572 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.835588 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.938663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.938710 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.938726 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.938745 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:13 crc kubenswrapper[4993]: I0129 12:06:13.938755 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:13Z","lastTransitionTime":"2026-01-29T12:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.042043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.042128 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.042153 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.042217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.042243 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.144986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.145035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.145043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.145059 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.145071 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.181462 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 19:55:02.308084785 +0000 UTC Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.189786 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.189935 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.189954 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.190031 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.190053 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.190178 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.248245 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.248303 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.248315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.248333 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.248345 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.351484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.351553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.351577 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.351604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.351625 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.454876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.454987 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.455005 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.455032 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.455074 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.558529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.558613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.558640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.558673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.558697 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.661261 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.661361 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.661402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.661434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.661474 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.763520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.763568 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.763581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.763600 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.763612 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.810772 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.810831 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.810855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.810881 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.810897 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.822032 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:14Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.826147 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.826208 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.826220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.826235 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.826245 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.837361 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:14Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.840692 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.840726 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.840735 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.840751 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.840761 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.851961 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:14Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.855168 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.855220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.855231 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.855247 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.855257 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.869926 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:14Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.874522 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.874661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.874750 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.874853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.874942 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.887155 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:14Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:14 crc kubenswrapper[4993]: E0129 12:06:14.887286 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.889746 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.889826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.889850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.889888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.889902 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.992434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.992464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.992475 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.992491 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:14 crc kubenswrapper[4993]: I0129 12:06:14.992503 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:14Z","lastTransitionTime":"2026-01-29T12:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.095144 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.095204 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.095217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.095236 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.095248 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.182452 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 12:35:59.864845252 +0000 UTC Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.190016 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:15 crc kubenswrapper[4993]: E0129 12:06:15.190231 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.197768 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.197815 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.197827 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.197843 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.197858 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.300226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.300289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.300305 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.300326 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.300342 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.403161 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.403278 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.403303 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.403327 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.403344 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.506143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.506454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.506541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.506639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.506722 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.610680 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.610750 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.610769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.610794 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.610812 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.714114 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.714166 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.714176 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.714217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.714231 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.817562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.817754 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.817786 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.817809 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.817822 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.920772 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.920820 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.920834 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.920853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:15 crc kubenswrapper[4993]: I0129 12:06:15.920865 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:15Z","lastTransitionTime":"2026-01-29T12:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.023847 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.023905 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.023921 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.023941 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.023956 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.126540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.126606 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.126617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.126632 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.126646 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.183516 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 15:07:07.732169481 +0000 UTC Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.189978 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.190019 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.189990 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:16 crc kubenswrapper[4993]: E0129 12:06:16.190223 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:16 crc kubenswrapper[4993]: E0129 12:06:16.190294 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:16 crc kubenswrapper[4993]: E0129 12:06:16.190363 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.228769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.228816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.228828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.228853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.228864 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.331436 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.331491 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.331504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.331520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.331533 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.434140 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.434231 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.434246 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.434300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.434316 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.536456 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.536544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.536568 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.536592 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.536613 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.638717 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.638765 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.638778 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.638798 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.638811 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.741329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.741366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.741374 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.741389 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.741399 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.843379 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.843415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.843425 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.843440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.843450 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.945626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.945720 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.945743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.945784 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:16 crc kubenswrapper[4993]: I0129 12:06:16.945872 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:16Z","lastTransitionTime":"2026-01-29T12:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.048671 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.048712 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.048721 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.048738 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.048748 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.151849 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.152377 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.152396 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.152415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.152429 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.184235 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 23:13:56.780127004 +0000 UTC Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.189557 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:17 crc kubenswrapper[4993]: E0129 12:06:17.189785 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.202936 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.215446 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.224737 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.247214 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.255452 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.255528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.255550 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.255581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.255603 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.257885 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.267436 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.280728 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.294350 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.308937 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.320786 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.330301 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.342455 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.353982 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.357690 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.357775 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.357793 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.357817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.357866 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.366587 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.377329 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.395871 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.410130 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:17Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.460359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.460417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.460432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.460454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.460471 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.562319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.562470 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.562484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.562500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.562511 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.665325 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.665412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.665440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.665468 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.665489 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.768486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.768524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.768532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.768549 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.768559 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.871195 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.871233 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.871242 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.871262 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.871274 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.973600 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.973646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.973658 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.973679 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:17 crc kubenswrapper[4993]: I0129 12:06:17.973692 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:17Z","lastTransitionTime":"2026-01-29T12:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.075963 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.075999 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.076007 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.076020 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.076029 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.178247 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.178293 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.178304 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.178319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.178329 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.184668 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 21:11:12.965160326 +0000 UTC Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.190054 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.190082 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.190087 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:18 crc kubenswrapper[4993]: E0129 12:06:18.190204 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:18 crc kubenswrapper[4993]: E0129 12:06:18.190351 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:18 crc kubenswrapper[4993]: E0129 12:06:18.190485 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.280530 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.280581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.280593 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.280608 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.280620 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.383124 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.383175 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.383209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.383230 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.383242 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.486432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.486493 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.486512 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.486531 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.486543 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.589065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.589139 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.589163 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.589229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.589247 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.692468 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.692521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.692532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.692547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.692558 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.795300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.795340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.795348 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.795361 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.795370 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.898280 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.898350 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.898366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.898389 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:18 crc kubenswrapper[4993]: I0129 12:06:18.898405 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:18Z","lastTransitionTime":"2026-01-29T12:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.001312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.001363 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.001372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.001390 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.001401 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.104741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.104795 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.104812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.104833 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.104851 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.185524 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 05:14:52.369829929 +0000 UTC Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.189937 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:19 crc kubenswrapper[4993]: E0129 12:06:19.190111 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.207275 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.207313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.207324 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.207340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.207355 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.309468 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.309526 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.309541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.309563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.309587 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.412435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.412469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.412503 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.412522 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.412533 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.514957 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.515070 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.515080 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.515100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.515116 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.618634 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.618839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.618926 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.619023 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.619114 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.721842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.722169 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.722326 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.722425 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.722517 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.825273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.825571 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.825693 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.825818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.825926 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.928731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.929146 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.929406 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.929579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:19 crc kubenswrapper[4993]: I0129 12:06:19.929732 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:19Z","lastTransitionTime":"2026-01-29T12:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.032760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.033021 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.033107 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.033217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.033559 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.136047 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.136417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.136619 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.136907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.137222 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.185990 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 23:48:11.679797286 +0000 UTC Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.190272 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:20 crc kubenswrapper[4993]: E0129 12:06:20.190569 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.190459 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:20 crc kubenswrapper[4993]: E0129 12:06:20.190854 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.190623 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:20 crc kubenswrapper[4993]: E0129 12:06:20.191255 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.240343 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.240396 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.240412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.240432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.240451 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.343008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.343053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.343065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.343084 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.343094 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.445700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.445731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.445739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.445751 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.445761 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.513127 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.524003 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.537237 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.547816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.547858 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.547867 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.547892 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.547911 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.549689 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.564864 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.585929 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da5d4789d4878d6d5533fcb535430ce3ff3f614b980e10ce217d2184545c3d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"message\\\":\\\".614022 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:05:50Z is after 2025-08-24T17:21:41Z]\\\\nI0129 12:05:50.614030 6388 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-b9gn8 in node crc\\\\nI0129 12:05:50.614032 6388 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI0129 12:05:50.614037 6388 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-bskq4\\\\nI0129 12:05:50.613999 6388 obj_retry.go:30\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.597860 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.610503 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.623479 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.635275 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.649790 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.652039 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.652088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.652098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.652112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.652121 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.660080 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.670179 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.681028 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.693004 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.711779 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.733762 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753686 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753746 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753762 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.753903 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:20Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.856119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.856174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.856232 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.856255 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.856272 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.959071 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.959101 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.959109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.959123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:20 crc kubenswrapper[4993]: I0129 12:06:20.959133 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:20Z","lastTransitionTime":"2026-01-29T12:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.061789 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.061838 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.061851 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.061868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.061880 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.164319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.164358 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.164367 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.164382 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.164391 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.186694 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 09:50:57.224087625 +0000 UTC Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.189662 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:21 crc kubenswrapper[4993]: E0129 12:06:21.189821 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.267278 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.267309 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.267318 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.267333 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.267343 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.369816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.369883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.369897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.369914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.369929 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.473089 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.473136 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.473144 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.473161 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.473171 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.575576 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.575618 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.575626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.575640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.575650 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.677737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.677772 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.677782 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.677798 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.677807 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.779994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.780040 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.780051 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.780067 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.780081 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.882526 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.882600 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.882661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.882695 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.882711 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.985748 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.985788 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.985797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.985817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:21 crc kubenswrapper[4993]: I0129 12:06:21.985829 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:21Z","lastTransitionTime":"2026-01-29T12:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.088160 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.088226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.088238 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.088254 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.088264 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.187715 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 22:58:37.626016661 +0000 UTC Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.189531 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.189581 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:22 crc kubenswrapper[4993]: E0129 12:06:22.189911 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190027 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:22 crc kubenswrapper[4993]: E0129 12:06:22.190037 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:22 crc kubenswrapper[4993]: E0129 12:06:22.190176 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190505 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190580 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.190589 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.293448 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.293523 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.293547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.293580 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.293604 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.396119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.396158 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.396167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.396198 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.396207 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.499077 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.499126 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.499171 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.499211 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.499227 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.602313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.602374 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.602398 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.602420 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.602435 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.704650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.704701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.704715 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.704732 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.704742 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.807310 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.807363 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.807378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.807402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.807415 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.909489 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.909521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.909531 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.909544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:22 crc kubenswrapper[4993]: I0129 12:06:22.909553 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:22Z","lastTransitionTime":"2026-01-29T12:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.011697 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.011783 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.011844 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.011874 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.011890 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.115134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.115178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.115203 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.115227 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.115239 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.188796 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 17:24:15.620213739 +0000 UTC Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.190150 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:23 crc kubenswrapper[4993]: E0129 12:06:23.190314 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.218402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.218435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.218444 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.218457 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.218466 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.321785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.321838 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.321850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.321868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.321880 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.423911 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.423973 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.423987 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.424030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.424044 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.527570 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.527614 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.527627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.527646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.527657 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.630353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.630391 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.630401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.630415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.630424 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.733482 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.733513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.733521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.733532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.733541 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.836868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.836936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.836961 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.836988 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.837010 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.939966 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.940027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.940036 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.940050 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:23 crc kubenswrapper[4993]: I0129 12:06:23.940060 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:23Z","lastTransitionTime":"2026-01-29T12:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.043597 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.043664 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.043691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.043738 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.043761 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.145955 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.146174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.146300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.146391 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.146480 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.189315 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 23:46:10.257476224 +0000 UTC Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.189412 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.189511 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.189888 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:24 crc kubenswrapper[4993]: E0129 12:06:24.190058 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:24 crc kubenswrapper[4993]: E0129 12:06:24.190245 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:24 crc kubenswrapper[4993]: E0129 12:06:24.190368 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.248602 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.248640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.248652 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.248665 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.248675 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.351494 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.351561 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.351579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.351614 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.351633 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.454012 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.454061 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.454073 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.454092 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.454104 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.558476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.558626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.558653 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.558690 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.558724 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.661069 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.661135 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.661149 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.661171 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.661203 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.763852 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.763905 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.763918 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.763943 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.763957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.866976 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.867015 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.867028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.867045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.867057 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.969915 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.969979 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.969994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.970013 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.970025 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.991789 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.991838 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.991851 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.991868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:24 crc kubenswrapper[4993]: I0129 12:06:24.991880 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:24Z","lastTransitionTime":"2026-01-29T12:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.006289 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.010065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.010113 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.010127 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.010147 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.010159 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.025791 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.030210 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.030266 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.030279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.030309 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.030322 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.042387 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.046817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.046862 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.046875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.046895 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.046908 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.058343 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.062044 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.062075 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.062085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.062100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.062113 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.075448 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.075634 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.077542 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.077578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.077586 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.077602 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.077613 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.181274 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.181333 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.181342 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.181363 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.181377 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.189681 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 13:31:02.223827423 +0000 UTC Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.189899 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.190053 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.231635 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.231790 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.231881 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:06:57.231863165 +0000 UTC m=+101.246993291 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.283891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.283946 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.283956 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.283969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.283977 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.386485 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.386557 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.386575 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.386599 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.386618 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.488882 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.488919 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.488928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.488943 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.488952 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.592691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.592757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.592770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.592790 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.592802 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.695432 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.695501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.695516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.695540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.695556 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.798312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.798368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.798382 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.798400 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.798415 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.901818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.901875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.901888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.901906 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.901918 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:25Z","lastTransitionTime":"2026-01-29T12:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.926883 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.927998 4993 scope.go:117] "RemoveContainer" containerID="e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9" Jan 29 12:06:25 crc kubenswrapper[4993]: E0129 12:06:25.928265 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.944804 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.963287 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:25 crc kubenswrapper[4993]: I0129 12:06:25.986492 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:25Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.003341 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.004266 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.004314 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.004332 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.004356 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.004373 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.022636 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.035511 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.055011 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.070066 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.091835 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.106830 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.106893 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.106914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.106939 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.106957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.121521 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.134469 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/0.log" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.134524 4993 generic.go:334] "Generic (PLEG): container finished" podID="d22ea793-b0f9-4b2c-8ea7-ed61728d900f" containerID="81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d" exitCode=1 Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.134556 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerDied","Data":"81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.134940 4993 scope.go:117] "RemoveContainer" containerID="81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.137901 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.150803 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.165827 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.178155 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.190410 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.190435 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.190476 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.190457 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 12:09:54.856530565 +0000 UTC Jan 29 12:06:26 crc kubenswrapper[4993]: E0129 12:06:26.190519 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:26 crc kubenswrapper[4993]: E0129 12:06:26.190654 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:26 crc kubenswrapper[4993]: E0129 12:06:26.190747 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.191857 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.205536 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.209102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.209140 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.209174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.209206 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.209218 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.217729 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.231432 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.244145 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.258494 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.270501 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.284901 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.296571 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.309021 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.311016 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.311046 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.311055 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.311069 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.311079 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.322486 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.339933 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.354772 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.366628 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.382396 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.398136 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.411501 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.413972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.414001 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.414009 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.414022 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.414030 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.424278 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.435420 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.452578 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:26Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.517429 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.517461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.517469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.517482 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.517491 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.619720 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.619759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.619792 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.619826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.619838 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.722748 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.723026 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.723227 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.723298 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.723412 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.825527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.825562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.825578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.825596 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.825607 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.928330 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.928385 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.928400 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.928417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:26 crc kubenswrapper[4993]: I0129 12:06:26.928429 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:26Z","lastTransitionTime":"2026-01-29T12:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.030680 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.030725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.030738 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.030755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.030766 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.133401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.133463 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.133477 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.133497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.133509 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.139659 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/0.log" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.139719 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerStarted","Data":"cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.153447 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.168834 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.181825 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.190195 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:27 crc kubenswrapper[4993]: E0129 12:06:27.190312 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.191157 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 10:43:58.995148468 +0000 UTC Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.196247 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.215885 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.225137 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.235924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.235960 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.235970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.235986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.235997 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.238308 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.250323 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.261256 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.272308 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.281673 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.292623 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.304705 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.315918 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.326028 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.335302 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.337792 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.337819 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.337827 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.337839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.337848 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.344572 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.355947 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.365596 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.378296 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.386702 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.397335 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.408585 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.419356 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.432412 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.439627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.439658 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.439666 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.439681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.439691 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.452658 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.465008 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.475922 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.489873 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.500947 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.512788 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.524707 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.535464 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.542326 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.542534 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.542684 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.542805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.542958 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.549739 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:27Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.645261 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.645322 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.645340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.645366 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.645383 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.749006 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.749321 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.749442 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.749522 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.749619 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.852277 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.852377 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.852391 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.852445 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.852457 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.955476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.955541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.955558 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.955577 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:27 crc kubenswrapper[4993]: I0129 12:06:27.955589 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:27Z","lastTransitionTime":"2026-01-29T12:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.058670 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.058986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.059065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.059137 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.059215 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.161973 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.162387 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.162551 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.162761 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.162898 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.189906 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.189932 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:28 crc kubenswrapper[4993]: E0129 12:06:28.190149 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:28 crc kubenswrapper[4993]: E0129 12:06:28.190223 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.189940 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:28 crc kubenswrapper[4993]: E0129 12:06:28.190301 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.191922 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 07:15:49.875002145 +0000 UTC Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.265611 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.265663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.265700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.265718 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.265730 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.368414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.368780 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.368902 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.369011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.369148 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.472233 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.472326 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.472347 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.472372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.472386 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.574943 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.574986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.574997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.575013 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.575024 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.677733 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.677782 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.677792 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.677808 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.677819 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.780488 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.780527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.780539 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.780554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.780564 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.882976 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.883005 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.883014 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.883027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.883036 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.985009 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.985078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.985092 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.985108 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:28 crc kubenswrapper[4993]: I0129 12:06:28.985118 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:28Z","lastTransitionTime":"2026-01-29T12:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.087375 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.087415 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.087428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.087445 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.087457 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.189701 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:29 crc kubenswrapper[4993]: E0129 12:06:29.189870 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.190501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.190536 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.190545 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.190557 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.190566 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.192251 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 01:49:40.146162224 +0000 UTC Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.292848 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.292905 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.292914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.292928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.292937 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.395476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.395518 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.395527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.395541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.395550 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.498133 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.498173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.498216 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.498232 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.498244 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.600691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.600722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.600730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.600743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.600753 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.703173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.703267 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.703283 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.703299 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.703337 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.805547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.805578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.805586 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.805601 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.805611 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.908621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.908659 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.908701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.908725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:29 crc kubenswrapper[4993]: I0129 12:06:29.908742 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:29Z","lastTransitionTime":"2026-01-29T12:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.010990 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.011030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.011040 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.011055 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.011065 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.113698 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.113735 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.113763 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.113777 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.113786 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.190025 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.190083 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:30 crc kubenswrapper[4993]: E0129 12:06:30.190164 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.190269 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:30 crc kubenswrapper[4993]: E0129 12:06:30.190283 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:30 crc kubenswrapper[4993]: E0129 12:06:30.190456 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.193292 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 13:14:39.507156093 +0000 UTC Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.217014 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.217057 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.217068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.217084 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.217094 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.320038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.320076 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.320085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.320100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.320108 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.422368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.422417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.422427 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.422442 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.422450 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.524272 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.524339 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.524357 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.524380 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.524398 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.626301 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.626349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.626358 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.626370 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.626379 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.728722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.728796 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.728817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.728845 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.728862 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.831220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.831267 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.831279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.831296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.831309 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.933499 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.933601 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.933616 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.933650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:30 crc kubenswrapper[4993]: I0129 12:06:30.933669 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:30Z","lastTransitionTime":"2026-01-29T12:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.035907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.035951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.035963 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.035980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.035992 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.138694 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.139264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.139296 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.139315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.139323 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.189780 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:31 crc kubenswrapper[4993]: E0129 12:06:31.189944 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.193476 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:44:33.156832747 +0000 UTC Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.241425 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.241484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.241493 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.241534 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.241544 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.344291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.344584 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.344669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.344757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.344837 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.447386 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.447425 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.447435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.447448 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.447457 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.550236 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.550470 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.550535 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.550630 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.550695 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.653034 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.653081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.653091 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.653105 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.653114 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.755237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.755491 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.755564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.755666 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.755743 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.857868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.858161 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.858269 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.858411 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.858492 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.960443 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.960686 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.960825 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.960961 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:31 crc kubenswrapper[4993]: I0129 12:06:31.961101 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:31Z","lastTransitionTime":"2026-01-29T12:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.063532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.063846 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.063916 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.063997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.064072 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.165928 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.166259 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.166444 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.166623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.166771 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.190424 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:32 crc kubenswrapper[4993]: E0129 12:06:32.190544 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.190446 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:32 crc kubenswrapper[4993]: E0129 12:06:32.190619 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.190424 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:32 crc kubenswrapper[4993]: E0129 12:06:32.190688 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.194518 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 09:59:31.01948085 +0000 UTC Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.269083 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.269143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.269158 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.269221 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.269238 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.371832 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.371885 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.371899 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.371913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.371922 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.473752 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.473796 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.473811 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.473830 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.473846 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.575779 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.575829 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.575870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.575888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.575901 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.677878 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.677942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.677961 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.677983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.678000 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.780885 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.780941 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.780957 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.780983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.781000 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.883983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.884033 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.884045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.884060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.884074 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.986899 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.986964 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.986974 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.986996 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:32 crc kubenswrapper[4993]: I0129 12:06:32.987015 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:32Z","lastTransitionTime":"2026-01-29T12:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.090120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.090167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.090204 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.090224 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.090238 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.190762 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:33 crc kubenswrapper[4993]: E0129 12:06:33.191030 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.192644 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.192688 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.192699 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.192725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.192741 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.195247 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 04:11:46.856810365 +0000 UTC Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.205166 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.295578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.295635 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.295649 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.295668 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.295679 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.398488 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.398519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.398528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.398540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.398550 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.501144 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.501264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.501284 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.501307 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.501324 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.603989 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.604035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.604047 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.604062 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.604073 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.706999 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.707072 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.707088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.707119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.707136 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.809927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.809969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.809977 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.809992 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.810001 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.913666 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.913730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.913739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.913755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:33 crc kubenswrapper[4993]: I0129 12:06:33.913765 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:33Z","lastTransitionTime":"2026-01-29T12:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.017262 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.017314 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.017327 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.017345 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.017357 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.120453 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.120492 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.120501 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.120515 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.120524 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.190390 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.190411 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:34 crc kubenswrapper[4993]: E0129 12:06:34.190547 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.190411 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:34 crc kubenswrapper[4993]: E0129 12:06:34.190714 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:34 crc kubenswrapper[4993]: E0129 12:06:34.190831 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.195524 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 21:14:25.795671135 +0000 UTC Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.222310 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.222345 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.222353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.222365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.222374 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.324737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.324805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.324826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.324851 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.324867 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.427150 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.427221 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.427235 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.427252 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.427262 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.529427 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.529516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.529535 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.529553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.529564 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.631688 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.631743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.631754 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.631769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.631779 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.733983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.734044 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.734065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.734095 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.734119 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.837088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.837131 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.837143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.837162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.837173 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.940521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.940563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.940576 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.940604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:34 crc kubenswrapper[4993]: I0129 12:06:34.940617 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:34Z","lastTransitionTime":"2026-01-29T12:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.043527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.043614 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.043634 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.043656 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.043673 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.139709 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.139762 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.139771 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.139782 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.139790 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.150596 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:35Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.153921 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.153986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.154001 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.154021 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.154033 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.167580 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:35Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.170840 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.170877 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.170885 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.170899 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.170908 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.181623 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:35Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.184713 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.184747 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.184757 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.184773 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.184783 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.189827 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.189949 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.195686 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 16:08:07.648003003 +0000 UTC Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.195680 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:35Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.199056 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.199117 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.199128 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.199143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.199152 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.210944 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:35Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:35 crc kubenswrapper[4993]: E0129 12:06:35.211071 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.212583 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.212619 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.212627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.212640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.212649 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.315134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.315209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.315228 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.315250 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.315266 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.418116 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.418159 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.418205 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.418237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.418254 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.521359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.521409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.521420 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.521442 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.521458 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.624027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.624075 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.624098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.624119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.624134 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.727315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.727381 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.727397 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.727421 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.727438 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.831041 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.831099 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.831112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.831131 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.831144 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.934793 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.935137 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.935439 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.935671 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:35 crc kubenswrapper[4993]: I0129 12:06:35.935852 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:35Z","lastTransitionTime":"2026-01-29T12:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.044306 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.044354 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.044365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.044380 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.044390 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.146785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.146832 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.146844 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.146862 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.146875 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.190159 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.190170 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:36 crc kubenswrapper[4993]: E0129 12:06:36.190326 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:36 crc kubenswrapper[4993]: E0129 12:06:36.190361 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.190778 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:36 crc kubenswrapper[4993]: E0129 12:06:36.191018 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.196566 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:39:35.01333635 +0000 UTC Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.248705 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.248991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.249138 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.249332 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.249476 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.352134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.352212 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.352237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.352257 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.352271 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.454913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.454955 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.454963 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.454978 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.454988 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.557742 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.557805 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.557823 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.557847 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.557865 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.660884 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.660939 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.660953 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.660972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.660986 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.763559 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.763610 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.763621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.763641 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.763655 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.865708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.865741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.865751 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.865767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.865778 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.968237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.968327 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.968354 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.968384 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:36 crc kubenswrapper[4993]: I0129 12:06:36.968403 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:36Z","lastTransitionTime":"2026-01-29T12:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.071907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.071971 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.071989 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.072013 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.072033 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.174362 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.174429 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.174449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.174476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.174499 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.190550 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:37 crc kubenswrapper[4993]: E0129 12:06:37.190864 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.196955 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 19:32:53.851470828 +0000 UTC Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.212106 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.232841 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.253284 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.273327 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.277708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.277740 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.277748 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.277761 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.278265 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.297688 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.310712 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.333749 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.359095 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.372331 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.380243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.380280 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.380289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.380302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.380311 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.383138 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.394629 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.406546 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.416299 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.426931 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.437763 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.447413 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.460066 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.471134 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:37Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.482784 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.482828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.482841 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.482860 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.482872 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.585350 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.585414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.585424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.585440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.585450 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.688120 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.688178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.688204 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.688234 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.688251 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.791405 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.792047 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.792100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.792137 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.792159 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.894975 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.895020 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.895028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.895043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.895054 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.997755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.997816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.997832 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.997854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:37 crc kubenswrapper[4993]: I0129 12:06:37.997867 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:37Z","lastTransitionTime":"2026-01-29T12:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.099854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.099918 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.099930 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.099947 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.099959 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.189825 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.189890 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:38 crc kubenswrapper[4993]: E0129 12:06:38.189945 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.189954 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:38 crc kubenswrapper[4993]: E0129 12:06:38.190046 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:38 crc kubenswrapper[4993]: E0129 12:06:38.190076 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.198094 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 09:34:37.078422831 +0000 UTC Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.201823 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.201858 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.201870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.201883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.201894 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.304855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.304933 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.304945 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.304962 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.304973 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.407834 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.407891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.407901 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.407914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.407923 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.510496 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.510549 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.510566 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.510590 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.510608 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.613909 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.613970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.613982 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.614005 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.614019 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.716669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.716750 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.716774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.716807 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.716832 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.820030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.820071 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.820080 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.820096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.820105 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.923062 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.923106 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.923118 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.923137 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:38 crc kubenswrapper[4993]: I0129 12:06:38.923149 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:38Z","lastTransitionTime":"2026-01-29T12:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.025519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.025560 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.025570 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.025590 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.025604 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.128446 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.128483 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.128490 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.128503 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.128511 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.190160 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:39 crc kubenswrapper[4993]: E0129 12:06:39.190358 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.191114 4993 scope.go:117] "RemoveContainer" containerID="e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.199255 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 11:57:13.01062422 +0000 UTC Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.231122 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.231201 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.231213 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.231229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.231240 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.333878 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.334614 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.334632 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.334659 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.334680 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.437755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.437802 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.437816 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.437843 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.437857 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.540554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.540595 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.540604 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.540623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.540634 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.643570 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.643628 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.643642 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.643661 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.643672 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.746488 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.746520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.746529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.746544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.746554 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.848639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.848673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.848681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.848693 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.848703 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.952038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.952084 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.952095 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.952112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:39 crc kubenswrapper[4993]: I0129 12:06:39.952125 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:39Z","lastTransitionTime":"2026-01-29T12:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.055087 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.055143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.055155 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.055178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.055224 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.157617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.157684 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.157703 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.157732 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.157764 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.183486 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/3.log" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.184151 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/2.log" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.187471 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" exitCode=1 Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.187539 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.187598 4993 scope.go:117] "RemoveContainer" containerID="e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.188580 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:06:40 crc kubenswrapper[4993]: E0129 12:06:40.188895 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.189496 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.189520 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.189499 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:40 crc kubenswrapper[4993]: E0129 12:06:40.189640 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:40 crc kubenswrapper[4993]: E0129 12:06:40.189738 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:40 crc kubenswrapper[4993]: E0129 12:06:40.189811 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.200314 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 06:30:55.027696074 +0000 UTC Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.205950 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.219031 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.231540 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.253222 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:40Z\\\",\\\"message\\\":\\\" openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021458 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:40.021493 7113 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021509 7113 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r in node crc\\\\nI0129 12:06:40.021519 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r after 0 failed attempt(s)\\\\nI0129 12:06:40.021527 7113 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.020755 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0129 12:06:40.021540 7113 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0129 12:06:40.020850 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.260297 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.260335 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.260346 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.260373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.260385 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.265094 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.275120 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.285909 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.301416 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.315027 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.329080 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.342180 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.353590 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.364108 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.364168 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.364201 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.364223 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.364237 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.370545 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.383716 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.397632 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.409568 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.422282 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.437237 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:40Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.467307 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.467358 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.467371 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.467387 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.467399 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.569830 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.569863 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.569876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.569890 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.569898 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.672104 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.672169 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.672179 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.672211 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.672222 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.775123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.775154 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.775162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.775175 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.775199 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.877032 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.877073 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.877085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.877099 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.877108 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.980379 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.980441 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.980465 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.980497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:40 crc kubenswrapper[4993]: I0129 12:06:40.980520 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:40Z","lastTransitionTime":"2026-01-29T12:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.082748 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.083011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.083121 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.083223 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.083306 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.186440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.186476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.186486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.186500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.186509 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.189533 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:41 crc kubenswrapper[4993]: E0129 12:06:41.189691 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.192284 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/3.log" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.201058 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 08:01:16.542025196 +0000 UTC Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.289209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.289478 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.289587 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.289659 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.289716 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.391970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.392022 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.392035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.392052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.392064 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.494696 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.494980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.495085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.495245 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.495347 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.597731 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.597762 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.597769 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.597800 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.597810 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.701075 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.701115 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.701123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.701136 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.701147 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.803549 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.803619 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.803636 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.803662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.803679 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.907349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.907395 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.907410 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.907428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:41 crc kubenswrapper[4993]: I0129 12:06:41.907443 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:41Z","lastTransitionTime":"2026-01-29T12:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.010681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.010746 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.010770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.010801 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.010825 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.114328 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.114391 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.114404 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.114423 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.114435 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.132141 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.132381 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.132352571 +0000 UTC m=+150.147482707 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.132433 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.132530 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.132592 4993 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.132643 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.132627899 +0000 UTC m=+150.147758035 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.132725 4993 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.132814 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.132797143 +0000 UTC m=+150.147927289 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.190244 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.190318 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.190433 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.190569 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.190745 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.190924 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.201886 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 12:51:24.549703356 +0000 UTC Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.217115 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.217167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.217178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.217220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.217233 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.233965 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.234026 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234158 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234178 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234215 4993 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234241 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234272 4993 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234288 4993 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234273 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.234256404 +0000 UTC m=+150.249386550 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:42 crc kubenswrapper[4993]: E0129 12:06:42.234356 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.234336876 +0000 UTC m=+150.249467022 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.320420 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.320458 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.320471 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.320487 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.320498 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.422477 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.422509 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.422517 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.422549 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.422559 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.524270 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.524338 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.524353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.524371 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.524383 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.626514 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.626553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.626563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.626578 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.626588 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.729940 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.729975 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.729985 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.729998 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.730009 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.832119 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.832157 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.832165 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.832193 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.832202 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.934456 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.934531 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.934544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.934564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:42 crc kubenswrapper[4993]: I0129 12:06:42.934602 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:42Z","lastTransitionTime":"2026-01-29T12:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.037598 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.037888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.037955 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.038024 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.038101 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.140111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.140146 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.140156 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.140171 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.140201 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.189924 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:43 crc kubenswrapper[4993]: E0129 12:06:43.190068 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.202391 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 11:50:54.650538083 +0000 UTC Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.242554 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.242594 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.242603 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.242620 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.242638 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.345445 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.345499 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.345510 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.345528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.345540 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.447924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.447972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.447986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.448007 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.448021 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.550439 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.550486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.550498 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.550514 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.550525 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.653282 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.653349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.653374 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.653404 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.653429 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.755693 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.755765 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.755785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.755808 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.755825 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.859080 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.859126 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.859143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.859166 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.859209 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.962232 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.962293 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.962312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.962335 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:43 crc kubenswrapper[4993]: I0129 12:06:43.962361 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:43Z","lastTransitionTime":"2026-01-29T12:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.065169 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.065263 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.065279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.065301 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.065317 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.167806 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.167844 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.167854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.167870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.167881 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.190498 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.190530 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.190541 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:44 crc kubenswrapper[4993]: E0129 12:06:44.190658 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:44 crc kubenswrapper[4993]: E0129 12:06:44.190731 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:44 crc kubenswrapper[4993]: E0129 12:06:44.190836 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.202923 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 14:41:49.289347519 +0000 UTC Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.270400 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.270452 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.270463 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.270479 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.270490 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.373337 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.373418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.373444 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.373486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.373512 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.475798 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.475837 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.475845 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.475859 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.475868 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.578508 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.578542 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.578550 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.578567 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.578579 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.681868 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.681918 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.681932 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.681951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.681965 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.785220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.785276 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.785293 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.785316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.785332 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.887922 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.887965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.887973 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.887988 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.887999 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.991024 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.991329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.991412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.991497 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:44 crc kubenswrapper[4993]: I0129 12:06:44.991599 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:44Z","lastTransitionTime":"2026-01-29T12:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.094872 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.095005 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.095027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.095053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.095069 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.190521 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.190727 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.196962 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.197034 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.197059 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.197088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.197111 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.203589 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 04:00:45.788002035 +0000 UTC Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.299505 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.299593 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.299606 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.299622 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.299661 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.402219 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.402506 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.402621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.402774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.402891 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.506670 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.506730 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.506747 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.506770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.506788 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.605959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.606011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.606028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.606060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.606078 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.624548 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.629646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.629689 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.629700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.629725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.629751 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.649034 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.653814 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.653890 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.653910 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.653942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.653961 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.672092 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.678595 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.678636 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.678646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.678664 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.678678 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.695598 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.699613 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.699657 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.699672 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.699691 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.699703 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.716431 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:45Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:45 crc kubenswrapper[4993]: E0129 12:06:45.716618 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.718110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.718148 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.718159 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.718176 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.718215 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.821068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.821313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.821418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.821492 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.821558 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.925207 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.925302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.925320 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.925373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:45 crc kubenswrapper[4993]: I0129 12:06:45.925390 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:45Z","lastTransitionTime":"2026-01-29T12:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.028541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.028592 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.028606 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.028625 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.028637 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.131171 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.131232 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.131241 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.131253 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.131262 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.189743 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:46 crc kubenswrapper[4993]: E0129 12:06:46.190163 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.189792 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:46 crc kubenswrapper[4993]: E0129 12:06:46.190453 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.189752 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:46 crc kubenswrapper[4993]: E0129 12:06:46.190709 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.204122 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 00:19:19.851367083 +0000 UTC Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.234057 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.234302 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.234368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.234449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.234556 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.337139 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.337431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.337598 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.337734 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.337866 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.441807 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.442111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.442226 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.442323 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.442392 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.545397 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.545669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.545732 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.545818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.545904 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.647664 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.647700 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.647710 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.647724 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.647737 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.749996 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.750028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.750038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.750052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.750062 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.851760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.851804 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.851815 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.851830 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.851842 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.953829 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.953859 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.953869 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.953883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:46 crc kubenswrapper[4993]: I0129 12:06:46.953893 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:46Z","lastTransitionTime":"2026-01-29T12:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.055965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.056003 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.056011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.056025 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.056033 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.158870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.159161 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.159279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.159396 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.159469 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.189887 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:47 crc kubenswrapper[4993]: E0129 12:06:47.190563 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.204323 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 21:41:29.431394916 +0000 UTC Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.209900 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.210127 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.224777 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.237336 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.247540 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.259321 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.261787 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.261817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.261826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.261839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.261847 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.269718 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.284071 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.307452 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e18b0b37e354e48bb4e453e7a6bf67eceb4febeda7fc088f6bc4262e72728dd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:11Z\\\",\\\"message\\\":\\\"1524 6699 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971547 6699 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.250:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de88cb48-af91-44f8-b3c0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:11.971620 6699 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:40Z\\\",\\\"message\\\":\\\" openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021458 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:40.021493 7113 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021509 7113 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r in node crc\\\\nI0129 12:06:40.021519 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r after 0 failed attempt(s)\\\\nI0129 12:06:40.021527 7113 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.020755 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0129 12:06:40.021540 7113 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0129 12:06:40.020850 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.317731 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.327467 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.340862 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.352834 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.363493 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.365284 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.365315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.365325 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.365340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.365350 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.375504 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.383886 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.395292 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.405869 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.416349 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:47Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.467068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.467110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.467123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.467138 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.467149 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.569384 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.569423 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.569434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.569450 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.569462 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.672649 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.672690 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.672702 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.672718 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.672730 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.775512 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.775547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.775555 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.775570 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.775581 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.878458 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.878853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.879035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.879259 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.879451 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.981945 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.981994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.982008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.982027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:47 crc kubenswrapper[4993]: I0129 12:06:47.982039 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:47Z","lastTransitionTime":"2026-01-29T12:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.084752 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.084799 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.084812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.084828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.084840 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.187265 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.187308 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.187323 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.187340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.187353 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.189716 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.189754 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:48 crc kubenswrapper[4993]: E0129 12:06:48.189819 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.189879 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:48 crc kubenswrapper[4993]: E0129 12:06:48.189917 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:48 crc kubenswrapper[4993]: E0129 12:06:48.190143 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.205346 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 00:18:48.36247803 +0000 UTC Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.291026 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.291095 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.291111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.291132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.291147 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.393662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.393701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.393708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.393725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.393734 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.496208 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.496248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.496256 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.496273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.496283 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.597871 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.597912 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.597920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.597935 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.597943 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.700573 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.700645 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.700673 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.700704 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.700725 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.802944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.803001 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.803015 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.803030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.803042 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.906419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.906474 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.906486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.906507 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:48 crc kubenswrapper[4993]: I0129 12:06:48.906521 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:48Z","lastTransitionTime":"2026-01-29T12:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.009172 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.009233 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.009248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.009265 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.009275 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.112764 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.112866 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.112881 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.112897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.112911 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.189864 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:49 crc kubenswrapper[4993]: E0129 12:06:49.190065 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.205832 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 03:01:30.949860243 +0000 UTC Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.215630 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.215674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.215683 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.215717 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.215727 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.318330 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.318569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.318637 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.318705 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.318762 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.421244 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.421281 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.421290 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.421304 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.421312 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.524412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.524480 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.524499 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.524517 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.524532 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.626520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.626562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.626574 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.626589 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.626600 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.729504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.729544 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.729553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.729569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.729580 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.831856 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.831912 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.831926 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.831942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.831954 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.934163 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.934228 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.934241 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.934257 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:49 crc kubenswrapper[4993]: I0129 12:06:49.934270 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:49Z","lastTransitionTime":"2026-01-29T12:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.036751 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.036813 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.036828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.036846 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.036857 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.139456 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.139500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.139510 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.139523 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.139538 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.190144 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.190215 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.190225 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:50 crc kubenswrapper[4993]: E0129 12:06:50.190310 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:50 crc kubenswrapper[4993]: E0129 12:06:50.190463 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:50 crc kubenswrapper[4993]: E0129 12:06:50.190499 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.207463 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 06:44:29.124720359 +0000 UTC Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.241903 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.241944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.241951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.241965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.241973 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.344042 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.344076 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.344085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.344100 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.344108 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.446322 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.446396 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.446408 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.446425 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.446438 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.549500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.549739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.549876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.549974 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.550061 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.652767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.652842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.652865 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.652930 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.652957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.755173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.755275 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.755291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.755315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.755332 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.857643 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.857674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.857681 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.857695 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.857703 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.961214 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.961272 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.961287 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.961311 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:50 crc kubenswrapper[4993]: I0129 12:06:50.961328 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:50Z","lastTransitionTime":"2026-01-29T12:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.063853 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.064109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.064231 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.064317 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.064445 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.167571 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.167626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.167637 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.167655 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.167666 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.190309 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:51 crc kubenswrapper[4993]: E0129 12:06:51.190882 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.207779 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 08:24:56.209434255 +0000 UTC Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.269958 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.270015 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.270027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.270045 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.270056 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.372533 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.372568 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.372582 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.372597 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.372606 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.475027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.475057 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.475068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.475084 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.475096 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.577038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.577076 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.577086 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.577102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.577112 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.679562 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.679636 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.679650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.679674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.679689 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.783126 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.783181 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.783216 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.783234 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.783245 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.885168 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.885243 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.885258 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.885277 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.885290 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.988451 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.988499 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.988510 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.988527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:51 crc kubenswrapper[4993]: I0129 12:06:51.988539 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:51Z","lastTransitionTime":"2026-01-29T12:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.091319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.091385 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.091412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.091434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.091452 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.190528 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:52 crc kubenswrapper[4993]: E0129 12:06:52.190664 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.190788 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.191227 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.191452 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:06:52 crc kubenswrapper[4993]: E0129 12:06:52.191464 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:52 crc kubenswrapper[4993]: E0129 12:06:52.191589 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:52 crc kubenswrapper[4993]: E0129 12:06:52.191603 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.193316 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.193359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.193373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.193393 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.193406 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.203816 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.208860 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 09:30:31.581696427 +0000 UTC Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.219221 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.229015 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.239353 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.249161 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.260466 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.269511 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.282594 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.295352 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.295449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.295460 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.295682 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.295693 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.299092 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:40Z\\\",\\\"message\\\":\\\" openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021458 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:40.021493 7113 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021509 7113 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r in node crc\\\\nI0129 12:06:40.021519 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r after 0 failed attempt(s)\\\\nI0129 12:06:40.021527 7113 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.020755 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0129 12:06:40.021540 7113 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0129 12:06:40.020850 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.311948 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.328773 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca036106-a4e3-43a1-995a-2e225174f440\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f34aad679fef25c08a240387412b0d423afcb66b6b50338926a355e1d2c2cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db7fb82b3a528a53375b9213a3af252b8e9fd80d23731c1660d08ce94e25749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda7a5c8db3f88f861b71392d31f5df467c7de45e03192201438899b6847799e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd3a2b8ef581a30ae5ad4678d5e2dcda7c53d96d219baa1363d5198baae9731\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fed706c97954479b34d7cef65cc6d0749bf8f064e8023b643899d3d84276da8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.339708 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.349351 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.359248 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.370642 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.378287 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.388816 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.397494 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.397528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.397538 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.397550 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.397559 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.399765 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.410931 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:52Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.499584 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.499621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.499632 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.499651 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.499664 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.602849 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.602883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.602892 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.602908 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.602919 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.709755 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.710023 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.710174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.710303 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.710465 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.813916 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.813970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.813980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.813997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.814006 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.917510 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.917563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.917579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.917701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:52 crc kubenswrapper[4993]: I0129 12:06:52.917725 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:52Z","lastTransitionTime":"2026-01-29T12:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.019525 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.019581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.019592 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.019607 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.019617 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.121822 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.122264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.122466 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.122660 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.123415 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.190160 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:53 crc kubenswrapper[4993]: E0129 12:06:53.190811 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.209168 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 03:15:15.019669685 +0000 UTC Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.225797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.225863 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.225876 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.225891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.225905 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.328291 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.328327 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.328337 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.328352 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.328363 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.430112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.430418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.430495 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.430583 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.430649 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.532873 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.532920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.532931 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.532947 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.532957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.634810 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.634846 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.634855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.634867 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.634875 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.743353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.743380 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.743388 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.743401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.743409 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.845914 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.846285 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.846431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.846609 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.846745 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.949059 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.949088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.949098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.949114 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:53 crc kubenswrapper[4993]: I0129 12:06:53.949124 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:53Z","lastTransitionTime":"2026-01-29T12:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.051461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.051505 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.051516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.051532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.051544 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.154239 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.154307 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.154334 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.154368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.154391 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.190258 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:54 crc kubenswrapper[4993]: E0129 12:06:54.190381 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.190449 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:54 crc kubenswrapper[4993]: E0129 12:06:54.190687 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.190866 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:54 crc kubenswrapper[4993]: E0129 12:06:54.191077 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.210757 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 12:02:06.780375904 +0000 UTC Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.257453 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.257771 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.257927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.258227 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.258437 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.361308 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.361344 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.361352 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.361365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.361375 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.464414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.464668 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.464793 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.464896 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.465037 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.567406 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.567687 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.567785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.567870 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.567954 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.670859 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.670917 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.670936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.670959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.670970 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.773365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.773424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.773435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.773457 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.773474 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.875824 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.875866 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.875875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.875891 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.875903 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.978904 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.978945 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.978957 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.978976 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:54 crc kubenswrapper[4993]: I0129 12:06:54.978988 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:54Z","lastTransitionTime":"2026-01-29T12:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.081478 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.081716 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.081787 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.081851 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.081909 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.184725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.184759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.184767 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.184781 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.184789 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.189667 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.189781 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.210938 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:04:54.807176049 +0000 UTC Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.287446 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.287506 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.287547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.287579 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.287602 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.390878 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.390936 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.390952 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.390978 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.390995 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.493944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.494004 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.494028 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.494054 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.494071 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.597550 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.597618 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.597639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.597670 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.597695 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.699773 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.699817 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.699826 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.699839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.699848 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.759098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.759320 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.759351 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.759439 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.759542 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.774524 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.778897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.778931 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.778940 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.778951 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.778960 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.791772 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.797622 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.797667 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.797679 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.797694 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.797703 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.812483 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.816176 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.816229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.816238 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.816252 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.816261 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.828112 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.831434 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.831466 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.831474 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.831486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.831495 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.845138 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:55Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.845345 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.846980 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.847030 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.847046 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.847067 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.847082 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.927091 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.927943 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:06:55 crc kubenswrapper[4993]: E0129 12:06:55.928117 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.950403 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.950440 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.950450 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.950461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:55 crc kubenswrapper[4993]: I0129 12:06:55.950469 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:55Z","lastTransitionTime":"2026-01-29T12:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.053556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.053598 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.053611 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.053627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.053638 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.156485 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.156552 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.156569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.156594 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.156613 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.190396 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.190568 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:56 crc kubenswrapper[4993]: E0129 12:06:56.190683 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.190734 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:56 crc kubenswrapper[4993]: E0129 12:06:56.190879 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:56 crc kubenswrapper[4993]: E0129 12:06:56.191029 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.211916 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 05:56:34.743924492 +0000 UTC Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.258937 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.258993 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.259006 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.259023 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.259033 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.362082 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.362132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.362144 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.362163 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.362197 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.465654 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.465701 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.465716 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.465735 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.465749 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.567993 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.568039 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.568049 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.568063 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.568073 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.674082 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.674155 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.674173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.674229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.674276 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.776986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.777027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.777035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.777049 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.777063 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.879656 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.879698 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.879708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.879722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.879733 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.981686 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.981712 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.981720 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.981733 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:56 crc kubenswrapper[4993]: I0129 12:06:56.981742 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:56Z","lastTransitionTime":"2026-01-29T12:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.084110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.084178 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.084220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.084244 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.084259 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.187672 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.187795 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.187812 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.187837 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.187848 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.190033 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:57 crc kubenswrapper[4993]: E0129 12:06:57.190164 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.204008 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.212598 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 18:13:52.451934017 +0000 UTC Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.216068 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.228132 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.239686 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.248256 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.269580 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca036106-a4e3-43a1-995a-2e225174f440\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f34aad679fef25c08a240387412b0d423afcb66b6b50338926a355e1d2c2cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db7fb82b3a528a53375b9213a3af252b8e9fd80d23731c1660d08ce94e25749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda7a5c8db3f88f861b71392d31f5df467c7de45e03192201438899b6847799e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd3a2b8ef581a30ae5ad4678d5e2dcda7c53d96d219baa1363d5198baae9731\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fed706c97954479b34d7cef65cc6d0749bf8f064e8023b643899d3d84276da8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.287462 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:57 crc kubenswrapper[4993]: E0129 12:06:57.287598 4993 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:57 crc kubenswrapper[4993]: E0129 12:06:57.288012 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs podName:7838e5a2-ff4a-42b8-aeef-fefbe1176fc4 nodeName:}" failed. No retries permitted until 2026-01-29 12:08:01.287994242 +0000 UTC m=+165.303124368 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs") pod "network-metrics-daemon-vdgbx" (UID: "7838e5a2-ff4a-42b8-aeef-fefbe1176fc4") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.292125 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.298543 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.298626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.298640 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.298660 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.298672 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.311106 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.322351 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.333297 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.342267 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.353330 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.364952 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.376437 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.390412 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.401122 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.401174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.401206 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.401221 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.401230 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.415451 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:40Z\\\",\\\"message\\\":\\\" openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021458 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:40.021493 7113 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021509 7113 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r in node crc\\\\nI0129 12:06:40.021519 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r after 0 failed attempt(s)\\\\nI0129 12:06:40.021527 7113 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.020755 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0129 12:06:40.021540 7113 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0129 12:06:40.020850 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.426616 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.436692 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.450530 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:06:57Z is after 2025-08-24T17:21:41Z" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.504035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.504096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.504111 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.504130 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.504142 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.605949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.605986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.605997 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.606016 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.606028 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.709072 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.709124 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.709135 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.709151 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.709163 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.811587 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.811623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.811639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.811655 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.811666 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.914131 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.914181 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.914209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.914228 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:57 crc kubenswrapper[4993]: I0129 12:06:57.914241 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:57Z","lastTransitionTime":"2026-01-29T12:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.016797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.016838 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.016847 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.016865 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.016875 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.119381 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.119436 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.119450 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.119471 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.119486 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.189635 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.189681 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.189658 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:06:58 crc kubenswrapper[4993]: E0129 12:06:58.189914 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:06:58 crc kubenswrapper[4993]: E0129 12:06:58.190017 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:06:58 crc kubenswrapper[4993]: E0129 12:06:58.190132 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.212722 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 14:43:03.043758089 +0000 UTC Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.221577 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.221620 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.221631 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.221647 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.221658 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.324005 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.324054 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.324067 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.324081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.324093 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.426236 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.426292 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.426304 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.426322 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.426334 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.528897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.528948 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.528959 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.528991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.529003 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.631340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.631389 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.631399 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.631414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.631425 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.733635 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.733672 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.733692 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.733712 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.733723 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.836949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.836989 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.837000 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.837017 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.837029 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.939794 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.940208 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.940303 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.940407 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:58 crc kubenswrapper[4993]: I0129 12:06:58.940502 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:58Z","lastTransitionTime":"2026-01-29T12:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.042920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.042960 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.042969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.042983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.042992 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.145756 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.145792 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.145803 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.145819 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.145830 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.189751 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:06:59 crc kubenswrapper[4993]: E0129 12:06:59.189903 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.212879 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 20:31:02.143496576 +0000 UTC Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.248229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.248264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.248273 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.248287 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.248299 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.351422 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.351465 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.351477 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.351494 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.351505 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.453205 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.453460 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.453528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.453601 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.453677 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.556148 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.556210 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.556221 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.556238 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.556252 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.658869 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.659975 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.660285 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.660385 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.660464 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.763174 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.763484 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.763676 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.763862 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.764006 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.866409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.866666 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.866765 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.866872 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.866941 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.969917 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.969954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.969965 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.969983 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:06:59 crc kubenswrapper[4993]: I0129 12:06:59.969998 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:06:59Z","lastTransitionTime":"2026-01-29T12:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.072458 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.072496 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.072504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.072536 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.072548 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.175341 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.175399 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.175411 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.175436 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.175449 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.190389 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.190404 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.190439 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:00 crc kubenswrapper[4993]: E0129 12:07:00.191310 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:00 crc kubenswrapper[4993]: E0129 12:07:00.191006 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:00 crc kubenswrapper[4993]: E0129 12:07:00.191228 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.213070 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 14:21:05.917045131 +0000 UTC Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.277949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.278264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.278346 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.278412 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.278479 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.380596 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.381096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.381172 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.381315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.381407 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.490761 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.490960 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.490981 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.491006 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.491024 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.594564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.595084 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.595233 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.595372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.595437 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.698438 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.698476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.698488 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.698503 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.698513 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.800788 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.800872 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.800886 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.800902 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.800916 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.903893 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.903940 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.903950 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.903992 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:00 crc kubenswrapper[4993]: I0129 12:07:00.904005 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:00Z","lastTransitionTime":"2026-01-29T12:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.007374 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.007418 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.007428 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.007444 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.007455 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.110287 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.110324 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.110333 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.110349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.110363 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.190014 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:01 crc kubenswrapper[4993]: E0129 12:07:01.190172 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.212972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.213014 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.213029 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.213048 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.213063 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.213373 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 23:35:00.936291389 +0000 UTC Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.315875 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.315923 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.315932 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.315949 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.315962 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.419079 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.419154 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.419169 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.419218 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.419231 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.521592 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.521635 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.521644 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.521676 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.521689 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.623605 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.623646 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.623655 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.623669 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.623679 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.726333 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.726383 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.726393 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.726409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.726420 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.829325 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.829378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.829390 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.829406 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.829439 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.931164 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.931229 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.931242 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.931264 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:01 crc kubenswrapper[4993]: I0129 12:07:01.931276 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:01Z","lastTransitionTime":"2026-01-29T12:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.034092 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.034519 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.034566 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.034588 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.034605 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.136424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.136482 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.136500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.136520 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.136531 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.190231 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.190307 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:02 crc kubenswrapper[4993]: E0129 12:07:02.190404 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.190281 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:02 crc kubenswrapper[4993]: E0129 12:07:02.190723 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:02 crc kubenswrapper[4993]: E0129 12:07:02.190884 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.214421 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 19:08:31.26670536 +0000 UTC Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.239141 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.239207 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.239248 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.239277 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.239295 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.342469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.342505 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.342513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.342528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.342538 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.444279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.444311 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.444320 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.444405 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.444426 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.546603 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.546650 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.546663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.546682 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.546698 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.649378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.649449 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.649461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.649480 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.649491 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.751518 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.751559 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.751567 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.751581 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.751591 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.854065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.854123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.854134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.854151 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.854166 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.956239 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.956281 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.956289 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.956305 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:02 crc kubenswrapper[4993]: I0129 12:07:02.956317 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:02Z","lastTransitionTime":"2026-01-29T12:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.058675 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.058722 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.058732 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.058749 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.058760 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.160943 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.160989 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.161002 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.161025 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.161037 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.190551 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:03 crc kubenswrapper[4993]: E0129 12:07:03.190708 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.215620 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 05:45:03.929239824 +0000 UTC Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.263649 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.263707 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.263743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.263770 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.263789 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.366027 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.366060 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.366068 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.366080 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.366089 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.468917 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.469023 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.469035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.469055 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.469067 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.571476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.571531 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.571545 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.571565 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.571576 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.673922 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.673971 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.673985 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.674002 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.674015 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.776467 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.776527 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.776543 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.776565 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.776580 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.879070 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.879401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.879500 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.879591 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.879695 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.981524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.981569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.981577 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.981592 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:03 crc kubenswrapper[4993]: I0129 12:07:03.981601 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:03Z","lastTransitionTime":"2026-01-29T12:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.083743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.083774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.083783 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.083822 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.083834 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.186508 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.186553 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.186568 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.186588 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.186602 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.190506 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.190557 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.190655 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:04 crc kubenswrapper[4993]: E0129 12:07:04.190857 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:04 crc kubenswrapper[4993]: E0129 12:07:04.191011 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:04 crc kubenswrapper[4993]: E0129 12:07:04.191158 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.216282 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 21:38:09.838745851 +0000 UTC Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.289047 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.289085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.289096 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.289112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.289123 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.391886 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.391927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.391954 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.391970 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.391979 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.494486 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.494523 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.494532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.494547 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.494559 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.596728 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.596809 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.596822 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.596839 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.597170 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.699745 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.699776 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.699786 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.699801 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.699809 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.802038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.802073 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.802083 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.802097 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.802106 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.904540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.904623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.904635 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.904663 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:04 crc kubenswrapper[4993]: I0129 12:07:04.904676 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:04Z","lastTransitionTime":"2026-01-29T12:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.007085 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.007288 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.007312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.007335 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.007350 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.110464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.110513 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.110524 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.110541 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.110552 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.189778 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.189982 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.213089 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.213139 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.213148 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.213160 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.213169 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.216522 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 13:53:41.068696433 +0000 UTC Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.315694 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.315737 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.315746 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.315759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.315770 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.418127 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.418220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.418237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.418261 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.418277 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.521090 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.521125 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.521133 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.521146 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.521156 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.623976 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.624022 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.624038 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.624099 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.624117 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.726833 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.726901 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.726924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.726946 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.726963 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.830353 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.830431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.830454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.830485 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.830509 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.880290 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.880351 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.880368 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.880393 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.880410 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.896224 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:05Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.900300 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.900409 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.900444 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.900466 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.900483 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.915789 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:05Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.918836 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.918897 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.918909 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.918924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.918933 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.934607 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:05Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.938668 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.938712 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.938726 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.938739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.938749 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.952323 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:05Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.955790 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.955822 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.955850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.955866 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.955875 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.968983 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T12:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2ddad0c9-9bb7-4482-acb5-3fbc806f4a38\\\",\\\"systemUUID\\\":\\\"6d0a6411-7fd6-4d96-a862-4c3278683d9f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:05Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:05 crc kubenswrapper[4993]: E0129 12:07:05.969107 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.970904 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.970929 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.970938 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.970953 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:05 crc kubenswrapper[4993]: I0129 12:07:05.970962 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:05Z","lastTransitionTime":"2026-01-29T12:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.072655 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.072729 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.072739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.072759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.072771 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.174994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.175033 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.175041 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.175056 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.175065 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.190466 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.190508 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:06 crc kubenswrapper[4993]: E0129 12:07:06.190598 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:06 crc kubenswrapper[4993]: E0129 12:07:06.190683 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.190734 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:06 crc kubenswrapper[4993]: E0129 12:07:06.190787 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.217395 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:21:45.039801726 +0000 UTC Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.277858 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.277904 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.277913 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.277927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.277937 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.380081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.380134 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.380144 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.380162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.380175 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.482981 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.483029 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.483039 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.483053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.483063 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.584942 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.584992 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.585008 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.585024 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.585035 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.687849 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.687903 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.687916 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.687934 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.687945 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.790938 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.790984 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.790994 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.791011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.791022 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.893450 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.893516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.893529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.893556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.893587 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.996315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.996364 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.996382 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.996402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:06 crc kubenswrapper[4993]: I0129 12:07:06.996417 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:06Z","lastTransitionTime":"2026-01-29T12:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.099237 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.099284 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.099297 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.099315 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.099325 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.189971 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:07 crc kubenswrapper[4993]: E0129 12:07:07.190180 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.201923 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.201963 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.201971 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.201985 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.201993 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.208799 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.218364 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:41:25.541619364 +0000 UTC Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.222097 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911a5ac308469b713cbf20c36f2fcd537d9496516717d6e74c66537e3bfb9cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.233496 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.244728 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qn5kr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8bf1534-f3e0-4302-a00d-17c0721fd168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://328cd9434f170d4710cb8a9d3af791897b3f2a8ad06058dded15971bbcebf18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7ghg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:52Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qn5kr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.267231 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca036106-a4e3-43a1-995a-2e225174f440\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f34aad679fef25c08a240387412b0d423afcb66b6b50338926a355e1d2c2cc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5db7fb82b3a528a53375b9213a3af252b8e9fd80d23731c1660d08ce94e25749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda7a5c8db3f88f861b71392d31f5df467c7de45e03192201438899b6847799e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd3a2b8ef581a30ae5ad4678d5e2dcda7c53d96d219baa1363d5198baae9731\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fed706c97954479b34d7cef65cc6d0749bf8f064e8023b643899d3d84276da8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89b0b3c67b80332076f391aa4d2d4e4c2f188e3aafe8024921576d3a9246e5b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ac5a421d7620c107069c405949833a485dd6488b0d87e48a85f6c30279845bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23deac8e116b3f920464b6ca9431b95741bb4cdbc996956a36933ba23dde90a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.280818 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84322ce4-8377-423b-8c5d-cf2dd90ad0e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"le observer\\\\nW0129 12:05:38.174673 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 12:05:38.174805 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 12:05:38.175424 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2032424861/tls.crt::/tmp/serving-cert-2032424861/tls.key\\\\\\\"\\\\nI0129 12:05:38.552383 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 12:05:38.554834 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 12:05:38.554854 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 12:05:38.554874 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 12:05:38.554881 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 12:05:38.563347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0129 12:05:38.563369 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563374 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 12:05:38.563379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 12:05:38.563385 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 12:05:38.563389 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 12:05:38.563392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0129 12:05:38.563506 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0129 12:05:38.564266 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.296450 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b9gn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22ea793-b0f9-4b2c-8ea7-ed61728d900f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:25Z\\\",\\\"message\\\":\\\"2026-01-29T12:05:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31\\\\n2026-01-29T12:05:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f015c872-844a-4d8d-adfd-3ae609442d31 to /host/opt/cni/bin/\\\\n2026-01-29T12:05:40Z [verbose] multus-daemon started\\\\n2026-01-29T12:05:40Z [verbose] Readiness Indicator file check\\\\n2026-01-29T12:06:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qljfq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b9gn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.304745 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.304802 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.304820 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.304842 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.304860 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.310428 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"930c2a97-0ce4-4b39-8ab6-b84b5b3add7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:06:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcc4447da286401f5611d7aca75a8eefab3a5d465b84d9261f36ad24556d491f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355786f4cae7eb3fe8a9303528a8abadf23bbb158d2b6b237af7d4441e6a11f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a43c3bcd0b20fc4013edc3f6334b47e5a16fea80708766e7c76f7e7706501b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1f4172a94670fa5d4c568776fbf9610ee1347f7dba8ea4d8b932a778ade5ea6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.321821 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adb1dafed2777475742940f202a7e9aa2a2ada056965c60dd10e62885629fd34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6453ed0ce018b6781c66128368ec5b977287897b19415cd39010736faeed05c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.332636 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clvrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4dec3d91-60d3-4d05-b1e3-dc7b76e3c2d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3611c2a8f15f1a98a78f639fc64f918455113757809bd654499f602162128ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9vfrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clvrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.348551 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73188108-8559-4fb4-929e-222f1b12acaa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://486f9e252c593511085731a26a9939af73cb28d15643421ccf50d86cb497ceaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fa426d92f042f2b4c863478775a5ae06478178248fc667170a1e947e0545854\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9e05c8ccaf4b62612f14031d73756d33d1d21dc5b9759a057e36ae149a74a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.362127 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.375939 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1df6750-623d-4dda-83cd-dec067f8fc1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67627441d3b9f1beaa3aca4b574ddf15cdcd0ea777cadfbe7a2b2959788e3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zlgfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bskq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.388638 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a56c564-0553-48b1-a375-c98f3901a0e2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d47ee52a8b1ad34e52218ddfafe81298730e89784880a195b17967ca8909c51c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e821a47fc9033e4bd354d48f704150d744a73f7605b392dd912a8559f5e8ecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a7111ce760ff1652473dd982daaeff9b04d889f760fb676d61a76371896cd83\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://513d5674046a16592be5f1743d87a4f45c991f62b23c19e0c69c8610c9abce14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10483a60c608761d8f68b3bb481b3aef43f7ad9859204187b10b71b303334303\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0a3cdee2a4d58fc284375ba005095527b66ef42765c782d0f713ae28a6bcf97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bcef45cab93cdcc04ac0c76950ecb39d0f6178efd5d6279d3690ab298fa4014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7sbzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f9pq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.407927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.407962 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.407972 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.407986 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.407998 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.408679 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6387d35-c6e4-42d4-9f89-6f2940878f8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T12:06:40Z\\\",\\\"message\\\":\\\" openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021458 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0129 12:06:40.021493 7113 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.021509 7113 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r in node crc\\\\nI0129 12:06:40.021519 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pzr6r after 0 failed attempt(s)\\\\nI0129 12:06:40.021527 7113 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-pzr6r\\\\nI0129 12:06:40.020755 7113 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-apiserver/kube-apiserver-crc after 0 failed attempt(s)\\\\nI0129 12:06:40.021540 7113 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0129 12:06:40.020850 7113 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T12:06:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5t8gs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzr6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.421521 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61f75260-59f9-40c7-868c-4802106efa65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://358ce7933db5e653d49e82f10c9066613e39c827d868d7a06c2de0e50a0488d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62fe2c43dcc5562b66d4d25937fcebc6018d5071b3769bcc516fbea187d1fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-shcp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-w74pd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.430936 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7dsr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vdgbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.442665 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"915b1e3d-f188-47c5-b3fc-9b83467838db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37cb49731bff551393b1d8ee48157eeab56f2a3643ebc3febb34b9c834d1bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a106983d710e5c28c07b19092b801fb9641a3e6d967c69f109663f706318aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T12:05:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T12:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T12:05:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.455060 4993 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T12:05:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca4dd7229b041445129ae550d7469102ddbb323f474de430ae9c3ad15d8773df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T12:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T12:07:07Z is after 2025-08-24T17:21:41Z" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.510993 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.511082 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.511095 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.511115 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.511127 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.613074 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.613123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.613135 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.613155 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.613195 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.716015 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.716053 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.716062 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.716075 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.716085 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.818759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.818799 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.818809 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.818823 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.818833 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.921132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.921176 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.921203 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.921218 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:07 crc kubenswrapper[4993]: I0129 12:07:07.921227 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:07Z","lastTransitionTime":"2026-01-29T12:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.023858 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.023907 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.023918 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.023935 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.023944 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.125766 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.125809 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.125818 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.125833 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.125842 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.190512 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.190538 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.191261 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:08 crc kubenswrapper[4993]: E0129 12:07:08.191371 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:08 crc kubenswrapper[4993]: E0129 12:07:08.191780 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.191816 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:07:08 crc kubenswrapper[4993]: E0129 12:07:08.191865 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:08 crc kubenswrapper[4993]: E0129 12:07:08.192172 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzr6r_openshift-ovn-kubernetes(c6387d35-c6e4-42d4-9f89-6f2940878f8a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.218893 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 20:03:44.983230828 +0000 UTC Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.227927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.227978 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.227991 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.228007 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.228020 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.330408 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.330463 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.330476 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.330494 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.330507 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.434056 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.434116 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.434130 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.434151 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.434165 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.538095 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.539022 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.539052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.539109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.539141 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.642658 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.642729 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.642750 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.642779 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.642800 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.745659 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.745702 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.745711 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.745727 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.745738 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.849290 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.849332 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.849340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.849356 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.849365 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.951563 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.951611 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.951623 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.951639 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:08 crc kubenswrapper[4993]: I0129 12:07:08.951650 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:08Z","lastTransitionTime":"2026-01-29T12:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.054295 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.054341 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.054351 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.054367 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.054378 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.157081 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.157132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.157143 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.157162 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.157174 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.189620 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:09 crc kubenswrapper[4993]: E0129 12:07:09.189812 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.220056 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 11:44:47.585737255 +0000 UTC Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.260462 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.260508 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.260521 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.260540 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.260551 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.362766 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.362810 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.362828 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.362850 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.362866 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.465536 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.465601 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.465630 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.465654 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.465669 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.568035 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.568094 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.568109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.568128 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.568142 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.670483 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.670528 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.670539 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.670558 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.670569 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.772703 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.772742 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.772759 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.772774 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.772784 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.874569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.874621 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.874638 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.874656 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.874668 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.977018 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.977066 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.977078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.977094 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:09 crc kubenswrapper[4993]: I0129 12:07:09.977106 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:09Z","lastTransitionTime":"2026-01-29T12:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.080083 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.080123 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.080132 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.080167 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.080179 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.182377 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.182424 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.182435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.182454 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.182467 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.190377 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:10 crc kubenswrapper[4993]: E0129 12:07:10.190517 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.190704 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:10 crc kubenswrapper[4993]: E0129 12:07:10.190764 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.190897 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:10 crc kubenswrapper[4993]: E0129 12:07:10.190963 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.220558 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 14:54:24.098567585 +0000 UTC Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.284161 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.284228 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.284238 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.284261 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.284281 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.387013 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.387065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.387073 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.387086 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.387094 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.489355 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.489394 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.489402 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.489416 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.489425 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.592725 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.592775 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.592784 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.592803 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.592813 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.696031 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.696088 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.696097 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.696109 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.696117 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.798452 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.798504 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.798516 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.798535 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.798548 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.900697 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.900765 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.900776 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.900791 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:10 crc kubenswrapper[4993]: I0129 12:07:10.900804 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:10Z","lastTransitionTime":"2026-01-29T12:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.004052 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.004101 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.004112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.004129 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.004140 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.106678 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.106728 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.106739 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.106756 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.106769 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.190744 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:11 crc kubenswrapper[4993]: E0129 12:07:11.190998 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.210110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.210206 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.210217 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.210241 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.210254 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.221157 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 08:09:04.011621236 +0000 UTC Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.314938 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.315046 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.315065 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.315097 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.315126 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.417469 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.417514 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.417525 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.417543 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.417555 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.520433 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.520498 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.520509 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.520529 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.520545 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.623306 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.623364 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.623376 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.623398 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.623414 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.726785 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.726844 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.726857 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.726879 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.726894 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.830932 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.831001 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.831011 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.831032 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.831042 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.934372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.934419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.934431 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.934448 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:11 crc kubenswrapper[4993]: I0129 12:07:11.934457 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:11Z","lastTransitionTime":"2026-01-29T12:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.037362 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.037408 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.037419 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.037435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.037447 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.139342 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.139390 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.139401 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.139417 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.139428 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.189952 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.190030 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:12 crc kubenswrapper[4993]: E0129 12:07:12.190138 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:12 crc kubenswrapper[4993]: E0129 12:07:12.190313 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.190385 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:12 crc kubenswrapper[4993]: E0129 12:07:12.190461 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.221841 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 20:16:56.434823771 +0000 UTC Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.241879 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.241920 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.241929 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.241944 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.241957 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.289620 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/1.log" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.290012 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/0.log" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.290056 4993 generic.go:334] "Generic (PLEG): container finished" podID="d22ea793-b0f9-4b2c-8ea7-ed61728d900f" containerID="cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b" exitCode=1 Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.290092 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerDied","Data":"cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.290125 4993 scope.go:117] "RemoveContainer" containerID="81b7d969d6b07bf63f7985436bebbb99cbfdf4672d04131da5d60b87affd196d" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.291018 4993 scope.go:117] "RemoveContainer" containerID="cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b" Jan 29 12:07:12 crc kubenswrapper[4993]: E0129 12:07:12.291279 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-b9gn8_openshift-multus(d22ea793-b0f9-4b2c-8ea7-ed61728d900f)\"" pod="openshift-multus/multus-b9gn8" podUID="d22ea793-b0f9-4b2c-8ea7-ed61728d900f" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.321717 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-f9pq9" podStartSLOduration=95.321697761 podStartE2EDuration="1m35.321697761s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.320099819 +0000 UTC m=+116.335229945" watchObservedRunningTime="2026-01-29 12:07:12.321697761 +0000 UTC m=+116.336827887" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.345247 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.345346 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.345359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.345378 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.345419 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.370464 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-w74pd" podStartSLOduration=94.37044594 podStartE2EDuration="1m34.37044594s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.358924015 +0000 UTC m=+116.374054161" watchObservedRunningTime="2026-01-29 12:07:12.37044594 +0000 UTC m=+116.385576066" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.382774 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=39.382752845 podStartE2EDuration="39.382752845s" podCreationTimestamp="2026-01-29 12:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.382449376 +0000 UTC m=+116.397579502" watchObservedRunningTime="2026-01-29 12:07:12.382752845 +0000 UTC m=+116.397882971" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.426147 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podStartSLOduration=95.426126171 podStartE2EDuration="1m35.426126171s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.413893278 +0000 UTC m=+116.429023404" watchObservedRunningTime="2026-01-29 12:07:12.426126171 +0000 UTC m=+116.441256297" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.447708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.447786 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.447797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.447819 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.447835 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.450803 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qn5kr" podStartSLOduration=94.450784752 podStartE2EDuration="1m34.450784752s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.449711944 +0000 UTC m=+116.464842090" watchObservedRunningTime="2026-01-29 12:07:12.450784752 +0000 UTC m=+116.465914878" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.471617 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=25.471597413 podStartE2EDuration="25.471597413s" podCreationTimestamp="2026-01-29 12:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.471225893 +0000 UTC m=+116.486356019" watchObservedRunningTime="2026-01-29 12:07:12.471597413 +0000 UTC m=+116.486727539" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.487065 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=93.487044091 podStartE2EDuration="1m33.487044091s" podCreationTimestamp="2026-01-29 12:05:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.48699383 +0000 UTC m=+116.502123976" watchObservedRunningTime="2026-01-29 12:07:12.487044091 +0000 UTC m=+116.502174217" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.515895 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=66.515880524 podStartE2EDuration="1m6.515880524s" podCreationTimestamp="2026-01-29 12:06:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.513822509 +0000 UTC m=+116.528952655" watchObservedRunningTime="2026-01-29 12:07:12.515880524 +0000 UTC m=+116.531010650" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.549457 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.549505 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.549515 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.549530 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.549540 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.573111 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=94.573092166 podStartE2EDuration="1m34.573092166s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.560246055 +0000 UTC m=+116.575376191" watchObservedRunningTime="2026-01-29 12:07:12.573092166 +0000 UTC m=+116.588222292" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.651854 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.651896 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.651910 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.651924 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.651933 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.754565 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.754612 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.754626 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.754641 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.754651 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.857489 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.857569 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.857582 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.857601 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.857614 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.960279 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.960319 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.960328 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.960342 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:12 crc kubenswrapper[4993]: I0129 12:07:12.960351 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:12Z","lastTransitionTime":"2026-01-29T12:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.062061 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.062102 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.062112 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.062126 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.062138 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.164043 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.164078 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.164090 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.164108 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.164119 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.189740 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:13 crc kubenswrapper[4993]: E0129 12:07:13.189896 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.222037 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 10:46:13.991777827 +0000 UTC Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.266556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.266605 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.266618 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.266636 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.266647 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.294750 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/1.log" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.369617 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.369662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.369674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.369693 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.369704 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.472219 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.472313 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.472329 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.472349 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.472362 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.575345 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.575393 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.575414 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.575433 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.575446 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.677328 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.677394 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.677411 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.677435 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.677454 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.779674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.779716 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.779729 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.779745 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.779756 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.882312 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.882360 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.882372 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.882387 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.882400 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.984662 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.984708 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.984719 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.984736 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:13 crc kubenswrapper[4993]: I0129 12:07:13.984746 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:13Z","lastTransitionTime":"2026-01-29T12:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.087468 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.087518 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.087532 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.087556 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.087572 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189360 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189395 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189431 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189407 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189475 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189493 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: E0129 12:07:14.189501 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189504 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.189395 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:14 crc kubenswrapper[4993]: E0129 12:07:14.189653 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:14 crc kubenswrapper[4993]: E0129 12:07:14.189667 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.223248 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 00:31:55.336818707 +0000 UTC Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.291952 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.292000 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.292016 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.292041 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.292057 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.394461 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.394508 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.394522 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.394539 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.394550 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.497277 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.497340 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.497352 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.497365 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.497375 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.599464 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.599667 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.599690 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.599719 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.599736 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.702674 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.702735 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.702743 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.702760 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.702769 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.806373 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.806430 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.806443 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.806462 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.806480 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.908969 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.909007 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.909017 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.909031 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:14 crc kubenswrapper[4993]: I0129 12:07:14.909041 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:14Z","lastTransitionTime":"2026-01-29T12:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.011981 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.012059 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.012074 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.012098 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.012118 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.114741 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.114786 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.114797 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.114813 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.114824 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.190086 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:15 crc kubenswrapper[4993]: E0129 12:07:15.190294 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.217400 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.217437 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.217448 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.217462 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.217472 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.223650 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 03:09:14.758455946 +0000 UTC Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.319888 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.319919 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.319927 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.319938 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.319946 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.422803 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.422877 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.422893 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.422922 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.422942 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.525110 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.525173 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.525221 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.525240 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.525253 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.627627 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.627676 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.627687 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.627704 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.627716 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.731762 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.731833 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.731855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.731883 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.731903 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.834801 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.834855 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.834867 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.834885 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.834897 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.937271 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.937309 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.937320 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.937359 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.937392 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.971564 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.972220 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.972238 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.972253 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 12:07:15 crc kubenswrapper[4993]: I0129 12:07:15.972264 4993 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T12:07:15Z","lastTransitionTime":"2026-01-29T12:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.013124 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-clvrz" podStartSLOduration=99.013108343 podStartE2EDuration="1m39.013108343s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:12.581267141 +0000 UTC m=+116.596397267" watchObservedRunningTime="2026-01-29 12:07:16.013108343 +0000 UTC m=+120.028238469" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.013577 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw"] Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.013896 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.016370 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.016381 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.016427 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.016583 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.079256 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.079328 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84d4e45a-4222-4d6e-91e1-bb2191db60c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.079357 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d4e45a-4222-4d6e-91e1-bb2191db60c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.079487 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d4e45a-4222-4d6e-91e1-bb2191db60c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.079540 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180671 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84d4e45a-4222-4d6e-91e1-bb2191db60c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180711 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d4e45a-4222-4d6e-91e1-bb2191db60c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180753 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d4e45a-4222-4d6e-91e1-bb2191db60c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180771 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180789 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.180836 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.181446 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84d4e45a-4222-4d6e-91e1-bb2191db60c5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.181721 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84d4e45a-4222-4d6e-91e1-bb2191db60c5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.187252 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84d4e45a-4222-4d6e-91e1-bb2191db60c5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.189551 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.189633 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:16 crc kubenswrapper[4993]: E0129 12:07:16.189698 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:16 crc kubenswrapper[4993]: E0129 12:07:16.189818 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.189938 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:16 crc kubenswrapper[4993]: E0129 12:07:16.190068 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.197847 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84d4e45a-4222-4d6e-91e1-bb2191db60c5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-899zw\" (UID: \"84d4e45a-4222-4d6e-91e1-bb2191db60c5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.224016 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 14:05:36.759246442 +0000 UTC Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.224069 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.232553 4993 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 12:07:16 crc kubenswrapper[4993]: I0129 12:07:16.327089 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" Jan 29 12:07:17 crc kubenswrapper[4993]: E0129 12:07:17.169870 4993 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 29 12:07:17 crc kubenswrapper[4993]: I0129 12:07:17.190888 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:17 crc kubenswrapper[4993]: E0129 12:07:17.191042 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:17 crc kubenswrapper[4993]: I0129 12:07:17.316823 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" event={"ID":"84d4e45a-4222-4d6e-91e1-bb2191db60c5","Type":"ContainerStarted","Data":"9183b18f9e132c8453bbbee1deec50060113b8c12234fc7f31b6408478d44c89"} Jan 29 12:07:17 crc kubenswrapper[4993]: I0129 12:07:17.316863 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" event={"ID":"84d4e45a-4222-4d6e-91e1-bb2191db60c5","Type":"ContainerStarted","Data":"b46278cbd53bfa4f70509472bb76434a64530adf24372fe6072e2e232c65655c"} Jan 29 12:07:17 crc kubenswrapper[4993]: I0129 12:07:17.332219 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-899zw" podStartSLOduration=100.332176476 podStartE2EDuration="1m40.332176476s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:17.331473627 +0000 UTC m=+121.346603753" watchObservedRunningTime="2026-01-29 12:07:17.332176476 +0000 UTC m=+121.347306612" Jan 29 12:07:18 crc kubenswrapper[4993]: E0129 12:07:18.016750 4993 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 12:07:18 crc kubenswrapper[4993]: I0129 12:07:18.190566 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:18 crc kubenswrapper[4993]: I0129 12:07:18.190590 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:18 crc kubenswrapper[4993]: E0129 12:07:18.191346 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:18 crc kubenswrapper[4993]: I0129 12:07:18.190626 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:18 crc kubenswrapper[4993]: E0129 12:07:18.191431 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:18 crc kubenswrapper[4993]: E0129 12:07:18.191553 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:19 crc kubenswrapper[4993]: I0129 12:07:19.190082 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:19 crc kubenswrapper[4993]: E0129 12:07:19.190675 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.190292 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.190366 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.190493 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:20 crc kubenswrapper[4993]: E0129 12:07:20.190609 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:20 crc kubenswrapper[4993]: E0129 12:07:20.190682 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:20 crc kubenswrapper[4993]: E0129 12:07:20.190934 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.191310 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.325792 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/3.log" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.328313 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerStarted","Data":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.329053 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:07:20 crc kubenswrapper[4993]: I0129 12:07:20.356017 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podStartSLOduration=103.355999505 podStartE2EDuration="1m43.355999505s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:20.355179934 +0000 UTC m=+124.370310080" watchObservedRunningTime="2026-01-29 12:07:20.355999505 +0000 UTC m=+124.371129651" Jan 29 12:07:21 crc kubenswrapper[4993]: I0129 12:07:21.157020 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vdgbx"] Jan 29 12:07:21 crc kubenswrapper[4993]: I0129 12:07:21.157135 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:21 crc kubenswrapper[4993]: E0129 12:07:21.157241 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:22 crc kubenswrapper[4993]: I0129 12:07:22.190460 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:22 crc kubenswrapper[4993]: I0129 12:07:22.190518 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:22 crc kubenswrapper[4993]: I0129 12:07:22.190462 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:22 crc kubenswrapper[4993]: E0129 12:07:22.190660 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:22 crc kubenswrapper[4993]: E0129 12:07:22.190815 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:22 crc kubenswrapper[4993]: E0129 12:07:22.190939 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:23 crc kubenswrapper[4993]: E0129 12:07:23.017656 4993 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 12:07:23 crc kubenswrapper[4993]: I0129 12:07:23.190314 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:23 crc kubenswrapper[4993]: E0129 12:07:23.190577 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:24 crc kubenswrapper[4993]: I0129 12:07:24.189451 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:24 crc kubenswrapper[4993]: I0129 12:07:24.189528 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:24 crc kubenswrapper[4993]: I0129 12:07:24.189461 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:24 crc kubenswrapper[4993]: E0129 12:07:24.189591 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:24 crc kubenswrapper[4993]: E0129 12:07:24.189708 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:24 crc kubenswrapper[4993]: E0129 12:07:24.189836 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:25 crc kubenswrapper[4993]: I0129 12:07:25.189672 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:25 crc kubenswrapper[4993]: I0129 12:07:25.190154 4993 scope.go:117] "RemoveContainer" containerID="cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b" Jan 29 12:07:25 crc kubenswrapper[4993]: E0129 12:07:25.190245 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:25 crc kubenswrapper[4993]: I0129 12:07:25.345939 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/1.log" Jan 29 12:07:25 crc kubenswrapper[4993]: I0129 12:07:25.346004 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerStarted","Data":"6d31403f624f5e15e97a01b0394e0dfbf2c956eebd89eee0e1fd00dcfde615b6"} Jan 29 12:07:25 crc kubenswrapper[4993]: I0129 12:07:25.941620 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:07:26 crc kubenswrapper[4993]: I0129 12:07:26.190164 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:26 crc kubenswrapper[4993]: E0129 12:07:26.190346 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 12:07:26 crc kubenswrapper[4993]: I0129 12:07:26.190396 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:26 crc kubenswrapper[4993]: I0129 12:07:26.190476 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:26 crc kubenswrapper[4993]: E0129 12:07:26.190612 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 12:07:26 crc kubenswrapper[4993]: E0129 12:07:26.190837 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 12:07:26 crc kubenswrapper[4993]: I0129 12:07:26.363929 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-b9gn8" podStartSLOduration=109.363910652 podStartE2EDuration="1m49.363910652s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:26.363602604 +0000 UTC m=+130.378732750" watchObservedRunningTime="2026-01-29 12:07:26.363910652 +0000 UTC m=+130.379040798" Jan 29 12:07:27 crc kubenswrapper[4993]: I0129 12:07:27.189751 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:27 crc kubenswrapper[4993]: E0129 12:07:27.190989 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vdgbx" podUID="7838e5a2-ff4a-42b8-aeef-fefbe1176fc4" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.189795 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.190167 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.190506 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.192482 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.193563 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.193565 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 12:07:28 crc kubenswrapper[4993]: I0129 12:07:28.194422 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 12:07:29 crc kubenswrapper[4993]: I0129 12:07:29.189900 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:07:29 crc kubenswrapper[4993]: I0129 12:07:29.196528 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 12:07:29 crc kubenswrapper[4993]: I0129 12:07:29.196930 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.269209 4993 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.309986 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.310484 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.310809 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.311244 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.311551 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-c8dv2"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.311976 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.312676 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wfg4q"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.313014 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.313618 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.313949 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.330242 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kc9dp"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.330711 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.330954 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.331141 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.333370 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.334563 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.336951 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.338833 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339200 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339293 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339390 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339470 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339561 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339651 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339740 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339825 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339907 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.339986 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340059 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340168 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340276 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340352 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340432 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340509 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340527 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340593 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340696 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.341164 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.341278 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.341357 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.341482 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.340423 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.345398 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.345567 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.347530 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.348908 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.349375 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.349587 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.349691 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.349928 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.349961 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.350215 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.350342 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.350472 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.351446 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.352739 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klxb6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.353307 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.354605 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.354708 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bn8qt"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355048 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355145 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355299 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355310 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355378 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355667 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355693 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355867 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.355944 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356005 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356221 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356418 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356449 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356537 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356424 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356718 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.356957 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.357411 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.357668 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.357772 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.357881 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.358158 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.358324 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.358336 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x7jkw"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.358370 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.365363 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.366349 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.366493 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.366601 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.366849 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367065 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mwxh6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367321 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367335 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367556 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367558 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-llk59"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367630 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367826 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367870 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.367963 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-s4sgr"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.368019 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.368306 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.370386 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.370958 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.371237 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.372120 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.374790 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.375926 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.396951 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.397195 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.397605 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.397838 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.407689 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.407946 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.408130 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.408458 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.409608 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.409780 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.411411 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.411693 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.414495 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.415016 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.415251 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.415844 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416250 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416315 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416354 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-images\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416394 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416443 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-config\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416476 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416503 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxjv\" (UniqueName: \"kubernetes.io/projected/4598a597-b97b-4461-a1ea-05c049d93ebe-kube-api-access-szxjv\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416524 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szjgz\" (UniqueName: \"kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416557 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq5cz\" (UniqueName: \"kubernetes.io/projected/a87d6206-4452-4dbc-a288-85c2a7b6332b-kube-api-access-gq5cz\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416604 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-config\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416637 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416671 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j25q\" (UniqueName: \"kubernetes.io/projected/a7656713-548a-4e6b-aba3-2ca0f87d942f-kube-api-access-8j25q\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416696 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ef5ad33-4ec5-4674-a517-8ade25ce2569-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416721 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d6206-4452-4dbc-a288-85c2a7b6332b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416744 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d6206-4452-4dbc-a288-85c2a7b6332b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416783 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2s48\" (UniqueName: \"kubernetes.io/projected/3ef5ad33-4ec5-4674-a517-8ade25ce2569-kube-api-access-l2s48\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416808 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416831 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-auth-proxy-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416854 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-service-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416888 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a7656713-548a-4e6b-aba3-2ca0f87d942f-machine-approver-tls\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416910 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598a597-b97b-4461-a1ea-05c049d93ebe-serving-cert\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416325 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.417171 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416384 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416450 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.417432 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416524 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.418006 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.418077 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.418811 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.419136 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.420347 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416598 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.416744 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421059 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421297 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421507 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421553 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421648 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421678 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421513 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.421938 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.424301 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.424852 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.424950 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.425052 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.425330 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.425831 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.427812 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.432246 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.433817 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.436548 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-79rwf"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.437070 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.441067 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.441504 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.441582 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.442938 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.443157 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-np8cv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.443702 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.443923 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.445396 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.445667 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.446469 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.447143 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.448008 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.449373 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.450028 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.451006 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.451051 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.459538 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.465743 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.470322 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.474035 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.474348 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.475142 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.477500 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.478869 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.491396 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.491828 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.492057 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.492525 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.492838 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbcsz"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.492965 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493252 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493275 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-c8dv2"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493284 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493294 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493301 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493303 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493448 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kc9dp"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493461 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klxb6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493474 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.493622 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mwxh6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.494884 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.496713 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.497812 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.498963 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.500040 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.501061 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.502198 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bn8qt"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.503224 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-79rwf"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.506372 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x7jkw"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.507938 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.509647 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-llk59"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.511648 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.512406 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.513351 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.514759 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wfg4q"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.516112 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.517789 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518300 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-auth-proxy-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518335 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwwv6\" (UniqueName: \"kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518356 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njpwl\" (UniqueName: \"kubernetes.io/projected/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-kube-api-access-njpwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518376 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-service-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518392 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518410 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518426 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaafca3f-0ac4-41c9-bc02-66fefa875aea-config\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518442 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a7656713-548a-4e6b-aba3-2ca0f87d942f-machine-approver-tls\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518457 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598a597-b97b-4461-a1ea-05c049d93ebe-serving-cert\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518473 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b20eefe-0735-4da1-9f09-0e455686b21e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518489 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518518 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518535 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-images\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518551 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518567 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-config\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518583 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518599 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l2nf\" (UniqueName: \"kubernetes.io/projected/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-kube-api-access-8l2nf\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518617 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518634 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518651 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518667 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518684 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxjv\" (UniqueName: \"kubernetes.io/projected/4598a597-b97b-4461-a1ea-05c049d93ebe-kube-api-access-szxjv\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518700 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518717 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-trusted-ca\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518733 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qklx\" (UniqueName: \"kubernetes.io/projected/b1ec8778-2a97-490f-a822-15d8bd226643-kube-api-access-8qklx\") pod \"downloads-7954f5f757-mwxh6\" (UID: \"b1ec8778-2a97-490f-a822-15d8bd226643\") " pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518748 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518762 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518778 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaafca3f-0ac4-41c9-bc02-66fefa875aea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518801 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szjgz\" (UniqueName: \"kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518822 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518839 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518857 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq5cz\" (UniqueName: \"kubernetes.io/projected/a87d6206-4452-4dbc-a288-85c2a7b6332b-kube-api-access-gq5cz\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518877 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518900 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518922 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eaafca3f-0ac4-41c9-bc02-66fefa875aea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518939 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b20eefe-0735-4da1-9f09-0e455686b21e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518954 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518971 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.518985 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519001 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519033 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zmwg\" (UniqueName: \"kubernetes.io/projected/7b20eefe-0735-4da1-9f09-0e455686b21e-kube-api-access-4zmwg\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519048 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519065 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-config\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519081 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996g7\" (UniqueName: \"kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519097 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519112 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519128 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d6206-4452-4dbc-a288-85c2a7b6332b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519145 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d6206-4452-4dbc-a288-85c2a7b6332b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519159 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519174 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j25q\" (UniqueName: \"kubernetes.io/projected/a7656713-548a-4e6b-aba3-2ca0f87d942f-kube-api-access-8j25q\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519212 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ef5ad33-4ec5-4674-a517-8ade25ce2569-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519236 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519257 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519277 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-serving-cert\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.519299 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj7tv\" (UniqueName: \"kubernetes.io/projected/ea098b70-73ee-4ee9-9f09-aa42910e215c-kube-api-access-xj7tv\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520111 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520153 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2s48\" (UniqueName: \"kubernetes.io/projected/3ef5ad33-4ec5-4674-a517-8ade25ce2569-kube-api-access-l2s48\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520173 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520208 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520225 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520241 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520258 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea098b70-73ee-4ee9-9f09-aa42910e215c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520273 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-config\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520289 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520306 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520322 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6czdj\" (UniqueName: \"kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.520929 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-images\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.521797 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-auth-proxy-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.521813 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ef5ad33-4ec5-4674-a517-8ade25ce2569-config\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.522165 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.522219 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.522403 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-service-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.523108 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7656713-548a-4e6b-aba3-2ca0f87d942f-config\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.523652 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d6206-4452-4dbc-a288-85c2a7b6332b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.523884 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.523976 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.524039 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.524644 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a7656713-548a-4e6b-aba3-2ca0f87d942f-machine-approver-tls\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.524666 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598a597-b97b-4461-a1ea-05c049d93ebe-config\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.524761 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598a597-b97b-4461-a1ea-05c049d93ebe-serving-cert\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.525585 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.526819 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3ef5ad33-4ec5-4674-a517-8ade25ce2569-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.527369 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d6206-4452-4dbc-a288-85c2a7b6332b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.528761 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-np8cv"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.530500 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.531642 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.531934 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.533030 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.534211 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.535562 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8gjn"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.536664 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-v8dh8"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.537129 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.537525 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.537629 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.539258 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.540433 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbcsz"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.540817 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v8dh8"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.541791 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8gjn"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.548161 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bfzr2"] Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.549340 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.552890 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.572178 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.593280 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.612294 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620791 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620839 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620861 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-serving-cert\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620878 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj7tv\" (UniqueName: \"kubernetes.io/projected/ea098b70-73ee-4ee9-9f09-aa42910e215c-kube-api-access-xj7tv\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620923 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620940 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620955 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.620972 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621004 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea098b70-73ee-4ee9-9f09-aa42910e215c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621019 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-config\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621036 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6czdj\" (UniqueName: \"kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621072 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621091 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwwv6\" (UniqueName: \"kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621108 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njpwl\" (UniqueName: \"kubernetes.io/projected/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-kube-api-access-njpwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621124 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621153 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621172 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaafca3f-0ac4-41c9-bc02-66fefa875aea-config\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621208 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b20eefe-0735-4da1-9f09-0e455686b21e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621226 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621253 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621294 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l2nf\" (UniqueName: \"kubernetes.io/projected/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-kube-api-access-8l2nf\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621311 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621333 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621368 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621385 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621401 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaafca3f-0ac4-41c9-bc02-66fefa875aea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621422 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621458 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-trusted-ca\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621474 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qklx\" (UniqueName: \"kubernetes.io/projected/b1ec8778-2a97-490f-a822-15d8bd226643-kube-api-access-8qklx\") pod \"downloads-7954f5f757-mwxh6\" (UID: \"b1ec8778-2a97-490f-a822-15d8bd226643\") " pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621490 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621527 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621552 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621580 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621613 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eaafca3f-0ac4-41c9-bc02-66fefa875aea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621631 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621649 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b20eefe-0735-4da1-9f09-0e455686b21e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621682 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621700 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621716 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621732 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621764 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621774 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621808 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zmwg\" (UniqueName: \"kubernetes.io/projected/7b20eefe-0735-4da1-9f09-0e455686b21e-kube-api-access-4zmwg\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621858 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996g7\" (UniqueName: \"kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621881 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621898 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.621983 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-config\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.624606 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.622764 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.622868 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.623325 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625013 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.624027 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.624099 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625198 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.623893 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625203 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.622432 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7b20eefe-0735-4da1-9f09-0e455686b21e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625543 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-serving-cert\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625559 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-trusted-ca\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625620 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.625908 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.627116 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.627243 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ea098b70-73ee-4ee9-9f09-aa42910e215c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.627491 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.627570 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.628071 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.628835 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.630699 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.630775 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.632148 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.632413 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.635437 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.650639 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.650723 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b20eefe-0735-4da1-9f09-0e455686b21e-serving-cert\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.659962 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.664635 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.673153 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.693012 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.712258 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.732265 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.758173 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.772435 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.792652 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.812419 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.833143 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.852683 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.872529 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.892391 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.912674 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.933148 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.952741 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.954476 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.972983 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 12:07:36 crc kubenswrapper[4993]: I0129 12:07:36.992616 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.012274 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.032550 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.052890 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.061945 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.063026 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.072303 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.092575 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.113081 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.132965 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.154167 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.173625 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.193372 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.213257 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.232890 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.253645 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.273136 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.292932 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.296782 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaafca3f-0ac4-41c9-bc02-66fefa875aea-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.313468 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.323003 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaafca3f-0ac4-41c9-bc02-66fefa875aea-config\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.333306 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.353217 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.372162 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.413517 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.431169 4993 request.go:700] Waited for 1.004808316s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmcc-proxy-tls&limit=500&resourceVersion=0 Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.432658 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.472181 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.493278 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.513076 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.534103 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.554052 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.573420 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.593724 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.613910 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.633516 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.653453 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.672917 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.692911 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.712963 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.734153 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.753712 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.772146 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.792827 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.812776 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.832596 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.852607 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.873563 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.893002 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.913359 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.941017 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.952405 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.973024 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 12:07:37 crc kubenswrapper[4993]: I0129 12:07:37.994065 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.012963 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.033479 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.052677 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.072994 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.092396 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.113478 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.133045 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.153011 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.199107 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szjgz\" (UniqueName: \"kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz\") pod \"route-controller-manager-6576b87f9c-kqng8\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.213073 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq5cz\" (UniqueName: \"kubernetes.io/projected/a87d6206-4452-4dbc-a288-85c2a7b6332b-kube-api-access-gq5cz\") pod \"openshift-apiserver-operator-796bbdcf4f-h88fc\" (UID: \"a87d6206-4452-4dbc-a288-85c2a7b6332b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.229551 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxjv\" (UniqueName: \"kubernetes.io/projected/4598a597-b97b-4461-a1ea-05c049d93ebe-kube-api-access-szxjv\") pod \"authentication-operator-69f744f599-c8dv2\" (UID: \"4598a597-b97b-4461-a1ea-05c049d93ebe\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.249906 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2s48\" (UniqueName: \"kubernetes.io/projected/3ef5ad33-4ec5-4674-a517-8ade25ce2569-kube-api-access-l2s48\") pod \"machine-api-operator-5694c8668f-wfg4q\" (UID: \"3ef5ad33-4ec5-4674-a517-8ade25ce2569\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.266774 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.269156 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j25q\" (UniqueName: \"kubernetes.io/projected/a7656713-548a-4e6b-aba3-2ca0f87d942f-kube-api-access-8j25q\") pod \"machine-approver-56656f9798-q2w6x\" (UID: \"a7656713-548a-4e6b-aba3-2ca0f87d942f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.272097 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.292942 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.312650 4993 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.333135 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.352977 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.373762 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.404338 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.411974 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.432237 4993 request.go:700] Waited for 1.882622732s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.433800 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.454733 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.456113 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.475720 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.488302 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj7tv\" (UniqueName: \"kubernetes.io/projected/ea098b70-73ee-4ee9-9f09-aa42910e215c-kube-api-access-xj7tv\") pod \"cluster-samples-operator-665b6dd947-rxfxk\" (UID: \"ea098b70-73ee-4ee9-9f09-aa42910e215c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:38 crc kubenswrapper[4993]: W0129 12:07:38.495864 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7656713_548a_4e6b_aba3_2ca0f87d942f.slice/crio-b7b89de3263111e39205cd5c46ef676f77db8de4ea49438423455ad555435b01 WatchSource:0}: Error finding container b7b89de3263111e39205cd5c46ef676f77db8de4ea49438423455ad555435b01: Status 404 returned error can't find the container with id b7b89de3263111e39205cd5c46ef676f77db8de4ea49438423455ad555435b01 Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.501709 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:07:38 crc kubenswrapper[4993]: W0129 12:07:38.509020 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7559669_0e20_491a_95ec_3beacab1c1e1.slice/crio-4685306264dbc9f62ae1e7d3c33448c3fecbc4bde2fa144b4ec2db8e80b641ab WatchSource:0}: Error finding container 4685306264dbc9f62ae1e7d3c33448c3fecbc4bde2fa144b4ec2db8e80b641ab: Status 404 returned error can't find the container with id 4685306264dbc9f62ae1e7d3c33448c3fecbc4bde2fa144b4ec2db8e80b641ab Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.509924 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l2nf\" (UniqueName: \"kubernetes.io/projected/fd864d61-94e3-4ccf-b5cd-d0b42ca329a4-kube-api-access-8l2nf\") pod \"console-operator-58897d9998-klxb6\" (UID: \"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4\") " pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.517258 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.526676 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eaafca3f-0ac4-41c9-bc02-66fefa875aea-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rzb4l\" (UID: \"eaafca3f-0ac4-41c9-bc02-66fefa875aea\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.537724 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.552482 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6czdj\" (UniqueName: \"kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj\") pod \"oauth-openshift-558db77b4-9zsmx\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.571487 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4bnwv\" (UID: \"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.588860 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zmwg\" (UniqueName: \"kubernetes.io/projected/7b20eefe-0735-4da1-9f09-0e455686b21e-kube-api-access-4zmwg\") pod \"openshift-config-operator-7777fb866f-rm9xh\" (UID: \"7b20eefe-0735-4da1-9f09-0e455686b21e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.599929 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.603064 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.607795 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc"] Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.608021 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.610474 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qklx\" (UniqueName: \"kubernetes.io/projected/b1ec8778-2a97-490f-a822-15d8bd226643-kube-api-access-8qklx\") pod \"downloads-7954f5f757-mwxh6\" (UID: \"b1ec8778-2a97-490f-a822-15d8bd226643\") " pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.649165 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njpwl\" (UniqueName: \"kubernetes.io/projected/0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb-kube-api-access-njpwl\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzcct\" (UID: \"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.653299 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwwv6\" (UniqueName: \"kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6\") pod \"console-f9d7485db-pbdv4\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.656287 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.668307 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.669059 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996g7\" (UniqueName: \"kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7\") pod \"controller-manager-879f6c89f-md8vs\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.677013 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.683905 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:38 crc kubenswrapper[4993]: W0129 12:07:38.686960 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda87d6206_4452_4dbc_a288_85c2a7b6332b.slice/crio-c3a78d69ff8b1772be10b24eb8be18fdcf9509206323c2a5d497efb2aae08f42 WatchSource:0}: Error finding container c3a78d69ff8b1772be10b24eb8be18fdcf9509206323c2a5d497efb2aae08f42: Status 404 returned error can't find the container with id c3a78d69ff8b1772be10b24eb8be18fdcf9509206323c2a5d497efb2aae08f42 Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.689900 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.758260 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wfg4q"] Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760255 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760308 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-node-pullsecrets\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760333 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-client\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760351 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvp74\" (UniqueName: \"kubernetes.io/projected/52adf4fb-76ea-4ca3-ab23-80c486b2a555-kube-api-access-xvp74\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760375 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760394 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760408 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7253cd01-1ade-43ba-a122-e2ba0f094648-config\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760440 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760455 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-serving-cert\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760469 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760484 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760499 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcvww\" (UniqueName: \"kubernetes.io/projected/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-kube-api-access-vcvww\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760514 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b2535d-da68-4605-9c49-b8a3d530c094-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760532 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760545 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkht7\" (UniqueName: \"kubernetes.io/projected/0517259c-9701-45de-9b5e-a71fb348fc2b-kube-api-access-vkht7\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760561 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760578 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snvbf\" (UniqueName: \"kubernetes.io/projected/0c2d1054-4b74-4702-b696-7cf26d9ff465-kube-api-access-snvbf\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760595 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-dir\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760616 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7253cd01-1ade-43ba-a122-e2ba0f094648-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760637 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-service-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760661 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit-dir\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760678 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760692 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760706 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0517259c-9701-45de-9b5e-a71fb348fc2b-service-ca-bundle\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760721 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760719 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.760734 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-serving-cert\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.761159 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.761196 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-client\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: E0129 12:07:38.761348 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.26133614 +0000 UTC m=+143.276466266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763324 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a30945c5-89ca-4a8c-8568-863454de6d1f-proxy-tls\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763393 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874q9\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763409 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-image-import-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763443 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763517 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz2v5\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-kube-api-access-vz2v5\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763572 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hchrk\" (UniqueName: \"kubernetes.io/projected/42b2535d-da68-4605-9c49-b8a3d530c094-kube-api-access-hchrk\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763595 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763613 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b2535d-da68-4605-9c49-b8a3d530c094-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763630 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-default-certificate\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763651 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-encryption-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763667 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763738 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763772 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-images\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763795 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-encryption-config\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763924 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjv6l\" (UniqueName: \"kubernetes.io/projected/fff9aa84-5f59-46e3-a40b-4622a6031d58-kube-api-access-zjv6l\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.763985 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764002 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n6qx\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-kube-api-access-7n6qx\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764027 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764048 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-serving-cert\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764067 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-config\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764084 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fff9aa84-5f59-46e3-a40b-4622a6031d58-metrics-tls\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764103 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-stats-auth\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764358 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a30945c5-89ca-4a8c-8568-863454de6d1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764392 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxxmc\" (UniqueName: \"kubernetes.io/projected/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-kube-api-access-hxxmc\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764419 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c2d1054-4b74-4702-b696-7cf26d9ff465-proxy-tls\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764472 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rs8p\" (UniqueName: \"kubernetes.io/projected/a30945c5-89ca-4a8c-8568-863454de6d1f-kube-api-access-8rs8p\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764529 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764570 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7253cd01-1ade-43ba-a122-e2ba0f094648-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764597 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-policies\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764635 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764664 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-client\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764726 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.764845 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-metrics-certs\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.779388 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" Jan 29 12:07:38 crc kubenswrapper[4993]: W0129 12:07:38.786322 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ef5ad33_4ec5_4674_a517_8ade25ce2569.slice/crio-56858d58a43ab182f3f5491cb8259b7656ad7152a636b862b657b7d6bf1077cf WatchSource:0}: Error finding container 56858d58a43ab182f3f5491cb8259b7656ad7152a636b862b657b7d6bf1077cf: Status 404 returned error can't find the container with id 56858d58a43ab182f3f5491cb8259b7656ad7152a636b862b657b7d6bf1077cf Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.830288 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-c8dv2"] Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865323 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865500 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865542 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snvbf\" (UniqueName: \"kubernetes.io/projected/0c2d1054-4b74-4702-b696-7cf26d9ff465-kube-api-access-snvbf\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865558 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-dir\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865580 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9fd3151-6897-49e6-9c92-36b7dcb2953d-cert\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:38 crc kubenswrapper[4993]: E0129 12:07:38.865614 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.365586234 +0000 UTC m=+143.380716360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865669 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7253cd01-1ade-43ba-a122-e2ba0f094648-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865694 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-certs\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865722 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865746 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-service-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865761 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-mountpoint-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865787 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit-dir\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865815 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sm96\" (UniqueName: \"kubernetes.io/projected/a9fd3151-6897-49e6-9c92-36b7dcb2953d-kube-api-access-2sm96\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865837 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv9h8\" (UniqueName: \"kubernetes.io/projected/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-kube-api-access-xv9h8\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865861 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865886 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865924 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0517259c-9701-45de-9b5e-a71fb348fc2b-service-ca-bundle\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865944 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc29514c-0991-488e-be96-46a03c78f543-config-volume\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865968 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.865999 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-client\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866022 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-serving-cert\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866044 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866088 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a30945c5-89ca-4a8c-8568-863454de6d1f-proxy-tls\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866123 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874q9\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866144 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-image-import-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866175 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dppps\" (UniqueName: \"kubernetes.io/projected/fb8104b0-8d5e-42e4-8912-859771cba6da-kube-api-access-dppps\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866222 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866227 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9f938607-0656-46f5-860b-45e736c42c1b-tmpfs\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866269 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c9pv\" (UniqueName: \"kubernetes.io/projected/9f938607-0656-46f5-860b-45e736c42c1b-kube-api-access-9c9pv\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866289 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-config\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866307 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-csi-data-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866337 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866356 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz2v5\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-kube-api-access-vz2v5\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866375 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-srv-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866393 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hchrk\" (UniqueName: \"kubernetes.io/projected/42b2535d-da68-4605-9c49-b8a3d530c094-kube-api-access-hchrk\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866411 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866429 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866445 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-node-bootstrap-token\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866465 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b2535d-da68-4605-9c49-b8a3d530c094-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866482 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-encryption-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866498 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-default-certificate\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866515 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2becf916-8ac8-48ad-ba1c-e679036dd069-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866532 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sns44\" (UniqueName: \"kubernetes.io/projected/2becf916-8ac8-48ad-ba1c-e679036dd069-kube-api-access-sns44\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866547 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjp7t\" (UniqueName: \"kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866563 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g7s6\" (UniqueName: \"kubernetes.io/projected/c67ab6d8-23c2-49de-b1b1-52a253777d13-kube-api-access-5g7s6\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866589 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866606 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-socket-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866660 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-webhook-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866676 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zkrd\" (UniqueName: \"kubernetes.io/projected/b6b4beb5-c112-470d-9041-85f5ebd7cc42-kube-api-access-8zkrd\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866705 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866740 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-images\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866756 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866773 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-encryption-config\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866789 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48tdw\" (UniqueName: \"kubernetes.io/projected/498b1a02-6998-4001-962d-15a59163f760-kube-api-access-48tdw\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866804 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9zk6\" (UniqueName: \"kubernetes.io/projected/bc29514c-0991-488e-be96-46a03c78f543-kube-api-access-t9zk6\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866818 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wclwf\" (UniqueName: \"kubernetes.io/projected/749b8957-29e5-434e-b736-3fd60318cb49-kube-api-access-wclwf\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866854 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc29514c-0991-488e-be96-46a03c78f543-metrics-tls\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866876 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866901 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-registration-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866937 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjv6l\" (UniqueName: \"kubernetes.io/projected/fff9aa84-5f59-46e3-a40b-4622a6031d58-kube-api-access-zjv6l\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866971 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.866986 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n6qx\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-kube-api-access-7n6qx\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867002 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867016 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867041 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-serving-cert\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867057 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-config\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867072 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fff9aa84-5f59-46e3-a40b-4622a6031d58-metrics-tls\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867088 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-stats-auth\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867107 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c2d1054-4b74-4702-b696-7cf26d9ff465-proxy-tls\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867123 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a30945c5-89ca-4a8c-8568-863454de6d1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867139 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldfzh\" (UniqueName: \"kubernetes.io/projected/ee9f4583-0a73-4090-a3e6-c0dada3788a6-kube-api-access-ldfzh\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.867156 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxxmc\" (UniqueName: \"kubernetes.io/projected/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-kube-api-access-hxxmc\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.868926 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9f4583-0a73-4090-a3e6-c0dada3788a6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.868987 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zxbc\" (UniqueName: \"kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869015 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rs8p\" (UniqueName: \"kubernetes.io/projected/a30945c5-89ca-4a8c-8568-863454de6d1f-kube-api-access-8rs8p\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869049 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869065 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7253cd01-1ade-43ba-a122-e2ba0f094648-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869081 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-policies\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869110 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869127 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-client\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869150 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869227 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-metrics-certs\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869274 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869294 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-srv-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869333 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-profile-collector-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869351 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-node-pullsecrets\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869368 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-serving-cert\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869387 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869404 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-client\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869423 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvp74\" (UniqueName: \"kubernetes.io/projected/52adf4fb-76ea-4ca3-ab23-80c486b2a555-kube-api-access-xvp74\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869451 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869496 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869525 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7253cd01-1ade-43ba-a122-e2ba0f094648-config\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869553 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/498b1a02-6998-4001-962d-15a59163f760-signing-key\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869599 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-plugins-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869649 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869668 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/498b1a02-6998-4001-962d-15a59163f760-signing-cabundle\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869688 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcvww\" (UniqueName: \"kubernetes.io/projected/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-kube-api-access-vcvww\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869707 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-serving-cert\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869740 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869756 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-apiservice-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869775 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869805 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p7fp\" (UniqueName: \"kubernetes.io/projected/acc7a110-91b0-4662-bdd2-0f7c998bc2e4-kube-api-access-4p7fp\") pod \"migrator-59844c95c7-9zb66\" (UID: \"acc7a110-91b0-4662-bdd2-0f7c998bc2e4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869843 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b2535d-da68-4605-9c49-b8a3d530c094-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869912 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p4rz\" (UniqueName: \"kubernetes.io/projected/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-kube-api-access-5p4rz\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869931 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.869948 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkht7\" (UniqueName: \"kubernetes.io/projected/0517259c-9701-45de-9b5e-a71fb348fc2b-kube-api-access-vkht7\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.870157 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-dir\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.871427 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-audit-dir\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.872210 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0517259c-9701-45de-9b5e-a71fb348fc2b-service-ca-bundle\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.872901 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.873863 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-service-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: E0129 12:07:38.874902 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.374874343 +0000 UTC m=+143.390004459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.876322 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-config\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.876488 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b2535d-da68-4605-9c49-b8a3d530c094-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.878925 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-image-import-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.879655 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-node-pullsecrets\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.881548 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.884458 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7253cd01-1ade-43ba-a122-e2ba0f094648-config\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.885893 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.885915 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.886078 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-serving-cert\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.887069 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.888297 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.888329 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.888735 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-default-certificate\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.888754 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52adf4fb-76ea-4ca3-ab23-80c486b2a555-audit-policies\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.888876 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.889008 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.889550 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c2d1054-4b74-4702-b696-7cf26d9ff465-images\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.889861 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.890377 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-ca\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.891114 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.892398 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-etcd-client\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.895558 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.896115 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b2535d-da68-4605-9c49-b8a3d530c094-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.897329 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7253cd01-1ade-43ba-a122-e2ba0f094648-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.897764 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-metrics-certs\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: W0129 12:07:38.900491 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4598a597_b97b_4461_a1ea_05c049d93ebe.slice/crio-5ffcdec52360fa9ac1ff1156cdcc7830be1d6bfa39788f9e812fd52ed9a8c2bf WatchSource:0}: Error finding container 5ffcdec52360fa9ac1ff1156cdcc7830be1d6bfa39788f9e812fd52ed9a8c2bf: Status 404 returned error can't find the container with id 5ffcdec52360fa9ac1ff1156cdcc7830be1d6bfa39788f9e812fd52ed9a8c2bf Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.900533 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.900768 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c2d1054-4b74-4702-b696-7cf26d9ff465-proxy-tls\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.900972 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.901343 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-serving-cert\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.903269 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-etcd-client\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.904291 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/52adf4fb-76ea-4ca3-ab23-80c486b2a555-encryption-config\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.904530 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a30945c5-89ca-4a8c-8568-863454de6d1f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.904610 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fff9aa84-5f59-46e3-a40b-4622a6031d58-metrics-tls\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.904731 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a30945c5-89ca-4a8c-8568-863454de6d1f-proxy-tls\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.905252 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-serving-cert\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.907561 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-encryption-config\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.908366 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.910976 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-etcd-client\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.911883 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkht7\" (UniqueName: \"kubernetes.io/projected/0517259c-9701-45de-9b5e-a71fb348fc2b-kube-api-access-vkht7\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.913211 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0517259c-9701-45de-9b5e-a71fb348fc2b-stats-auth\") pod \"router-default-5444994796-s4sgr\" (UID: \"0517259c-9701-45de-9b5e-a71fb348fc2b\") " pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.921157 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh"] Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.971492 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.971885 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-profile-collector-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.971929 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-serving-cert\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.971975 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972003 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/498b1a02-6998-4001-962d-15a59163f760-signing-key\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972028 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-plugins-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972048 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/498b1a02-6998-4001-962d-15a59163f760-signing-cabundle\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972117 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-apiservice-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972142 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p7fp\" (UniqueName: \"kubernetes.io/projected/acc7a110-91b0-4662-bdd2-0f7c998bc2e4-kube-api-access-4p7fp\") pod \"migrator-59844c95c7-9zb66\" (UID: \"acc7a110-91b0-4662-bdd2-0f7c998bc2e4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972199 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p4rz\" (UniqueName: \"kubernetes.io/projected/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-kube-api-access-5p4rz\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972246 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9fd3151-6897-49e6-9c92-36b7dcb2953d-cert\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972268 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-certs\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972290 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972312 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-mountpoint-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972335 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sm96\" (UniqueName: \"kubernetes.io/projected/a9fd3151-6897-49e6-9c92-36b7dcb2953d-kube-api-access-2sm96\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972358 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv9h8\" (UniqueName: \"kubernetes.io/projected/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-kube-api-access-xv9h8\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972380 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc29514c-0991-488e-be96-46a03c78f543-config-volume\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972406 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dppps\" (UniqueName: \"kubernetes.io/projected/fb8104b0-8d5e-42e4-8912-859771cba6da-kube-api-access-dppps\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972424 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9f938607-0656-46f5-860b-45e736c42c1b-tmpfs\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972440 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c9pv\" (UniqueName: \"kubernetes.io/projected/9f938607-0656-46f5-860b-45e736c42c1b-kube-api-access-9c9pv\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972457 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-config\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972472 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-csi-data-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972501 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-srv-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972523 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972540 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-node-bootstrap-token\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972556 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g7s6\" (UniqueName: \"kubernetes.io/projected/c67ab6d8-23c2-49de-b1b1-52a253777d13-kube-api-access-5g7s6\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972574 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2becf916-8ac8-48ad-ba1c-e679036dd069-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972591 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sns44\" (UniqueName: \"kubernetes.io/projected/2becf916-8ac8-48ad-ba1c-e679036dd069-kube-api-access-sns44\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972607 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjp7t\" (UniqueName: \"kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972623 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-socket-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972647 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-webhook-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972663 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zkrd\" (UniqueName: \"kubernetes.io/projected/b6b4beb5-c112-470d-9041-85f5ebd7cc42-kube-api-access-8zkrd\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972684 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972700 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48tdw\" (UniqueName: \"kubernetes.io/projected/498b1a02-6998-4001-962d-15a59163f760-kube-api-access-48tdw\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972715 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9zk6\" (UniqueName: \"kubernetes.io/projected/bc29514c-0991-488e-be96-46a03c78f543-kube-api-access-t9zk6\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972732 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wclwf\" (UniqueName: \"kubernetes.io/projected/749b8957-29e5-434e-b736-3fd60318cb49-kube-api-access-wclwf\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972753 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc29514c-0991-488e-be96-46a03c78f543-metrics-tls\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972769 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972785 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-registration-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972815 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972848 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldfzh\" (UniqueName: \"kubernetes.io/projected/ee9f4583-0a73-4090-a3e6-c0dada3788a6-kube-api-access-ldfzh\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972867 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9f4583-0a73-4090-a3e6-c0dada3788a6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972888 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zxbc\" (UniqueName: \"kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.972931 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-srv-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:38 crc kubenswrapper[4993]: E0129 12:07:38.974327 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.474296357 +0000 UTC m=+143.489426493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.978158 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-csi-data-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.978354 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-mountpoint-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.978593 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-plugins-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.979224 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-socket-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.979422 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b6b4beb5-c112-470d-9041-85f5ebd7cc42-registration-dir\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:38 crc kubenswrapper[4993]: I0129 12:07:38.995812 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-klxb6"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.009605 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.073809 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.074197 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.574169554 +0000 UTC m=+143.589299680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.083759 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.084081 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.085239 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-srv-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.085321 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hchrk\" (UniqueName: \"kubernetes.io/projected/42b2535d-da68-4605-9c49-b8a3d530c094-kube-api-access-hchrk\") pod \"kube-storage-version-migrator-operator-b67b599dd-47m4j\" (UID: \"42b2535d-da68-4605-9c49-b8a3d530c094\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.086010 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2becf916-8ac8-48ad-ba1c-e679036dd069-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.086069 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcvww\" (UniqueName: \"kubernetes.io/projected/38b7a4c3-bead-47c0-a37a-b0bf359b20aa-kube-api-access-vcvww\") pod \"apiserver-76f77b778f-kc9dp\" (UID: \"38b7a4c3-bead-47c0-a37a-b0bf359b20aa\") " pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.086287 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9f4583-0a73-4090-a3e6-c0dada3788a6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.086801 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.086967 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rs8p\" (UniqueName: \"kubernetes.io/projected/a30945c5-89ca-4a8c-8568-863454de6d1f-kube-api-access-8rs8p\") pod \"machine-config-controller-84d6567774-2fbch\" (UID: \"a30945c5-89ca-4a8c-8568-863454de6d1f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.087685 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.087867 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9fd3151-6897-49e6-9c92-36b7dcb2953d-cert\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.088058 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/498b1a02-6998-4001-962d-15a59163f760-signing-cabundle\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.090326 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/498b1a02-6998-4001-962d-15a59163f760-signing-key\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.091688 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc29514c-0991-488e-be96-46a03c78f543-config-volume\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.092510 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-config\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.092899 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9f938607-0656-46f5-860b-45e736c42c1b-tmpfs\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.094301 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.094780 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-webhook-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.095055 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-serving-cert\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.095075 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb8104b0-8d5e-42e4-8912-859771cba6da-profile-collector-cert\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.095854 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f938607-0656-46f5-860b-45e736c42c1b-apiservice-cert\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.097448 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874q9\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.099130 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snvbf\" (UniqueName: \"kubernetes.io/projected/0c2d1054-4b74-4702-b696-7cf26d9ff465-kube-api-access-snvbf\") pod \"machine-config-operator-74547568cd-d8nzx\" (UID: \"0c2d1054-4b74-4702-b696-7cf26d9ff465\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.099284 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-certs\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.099918 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc29514c-0991-488e-be96-46a03c78f543-metrics-tls\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.100023 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd864d61_94e3_4ccf_b5cd_d0b42ca329a4.slice/crio-7ad35825266d59a3862893709fb3a620f230213ff16f6ef75bb9f168b018bd54 WatchSource:0}: Error finding container 7ad35825266d59a3862893709fb3a620f230213ff16f6ef75bb9f168b018bd54: Status 404 returned error can't find the container with id 7ad35825266d59a3862893709fb3a620f230213ff16f6ef75bb9f168b018bd54 Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.100606 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.100664 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/749b8957-29e5-434e-b736-3fd60318cb49-node-bootstrap-token\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.107055 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjv6l\" (UniqueName: \"kubernetes.io/projected/fff9aa84-5f59-46e3-a40b-4622a6031d58-kube-api-access-zjv6l\") pod \"dns-operator-744455d44c-bn8qt\" (UID: \"fff9aa84-5f59-46e3-a40b-4622a6031d58\") " pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.128790 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7253cd01-1ade-43ba-a122-e2ba0f094648-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xrmnd\" (UID: \"7253cd01-1ade-43ba-a122-e2ba0f094648\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.147982 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.154445 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.170986 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.172807 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz2v5\" (UniqueName: \"kubernetes.io/projected/9d86802e-e37a-42e9-aa5e-e0aff76ba6d7-kube-api-access-vz2v5\") pod \"ingress-operator-5b745b69d9-llk59\" (UID: \"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.172918 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-srv-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.173631 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n6qx\" (UniqueName: \"kubernetes.io/projected/fd0e07ac-1b26-4c71-9b89-9a19e45397ca-kube-api-access-7n6qx\") pod \"cluster-image-registry-operator-dc59b4c8b-2ztwc\" (UID: \"fd0e07ac-1b26-4c71-9b89-9a19e45397ca\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.173678 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c67ab6d8-23c2-49de-b1b1-52a253777d13-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.174410 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.174822 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.674809103 +0000 UTC m=+143.689939229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.181649 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.185383 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvp74\" (UniqueName: \"kubernetes.io/projected/52adf4fb-76ea-4ca3-ab23-80c486b2a555-kube-api-access-xvp74\") pod \"apiserver-7bbb656c7d-hjhzh\" (UID: \"52adf4fb-76ea-4ca3-ab23-80c486b2a555\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.188317 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.204705 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxxmc\" (UniqueName: \"kubernetes.io/projected/a0a5d6a8-59ec-4233-ae13-0404ea70c95c-kube-api-access-hxxmc\") pod \"etcd-operator-b45778765-x7jkw\" (UID: \"a0a5d6a8-59ec-4233-ae13-0404ea70c95c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.216871 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.221644 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.223692 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.227417 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dppps\" (UniqueName: \"kubernetes.io/projected/fb8104b0-8d5e-42e4-8912-859771cba6da-kube-api-access-dppps\") pod \"catalog-operator-68c6474976-27cpg\" (UID: \"fb8104b0-8d5e-42e4-8912-859771cba6da\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.238827 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.243940 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.252068 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod887d412a_3e41_473f_9954_4fcfccf6b2ea.slice/crio-1e6729faa1ae6b24e4f34106769ab1384e011f7e7d99fd214720307aeff6e84c WatchSource:0}: Error finding container 1e6729faa1ae6b24e4f34106769ab1384e011f7e7d99fd214720307aeff6e84c: Status 404 returned error can't find the container with id 1e6729faa1ae6b24e4f34106769ab1384e011f7e7d99fd214720307aeff6e84c Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.253679 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.256631 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c9pv\" (UniqueName: \"kubernetes.io/projected/9f938607-0656-46f5-860b-45e736c42c1b-kube-api-access-9c9pv\") pod \"packageserver-d55dfcdfc-7nkd9\" (UID: \"9f938607-0656-46f5-860b-45e736c42c1b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.261362 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.270449 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p7fp\" (UniqueName: \"kubernetes.io/projected/acc7a110-91b0-4662-bdd2-0f7c998bc2e4-kube-api-access-4p7fp\") pod \"migrator-59844c95c7-9zb66\" (UID: \"acc7a110-91b0-4662-bdd2-0f7c998bc2e4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.275633 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.275961 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.775950017 +0000 UTC m=+143.791080143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.289110 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aa2a004_0aa7_499e_b2d6_6049dfbd9ddb.slice/crio-68e3b0d2c7b5a8bae587797f20786e404c6a71d9f357b2052068f0a37c10cf67 WatchSource:0}: Error finding container 68e3b0d2c7b5a8bae587797f20786e404c6a71d9f357b2052068f0a37c10cf67: Status 404 returned error can't find the container with id 68e3b0d2c7b5a8bae587797f20786e404c6a71d9f357b2052068f0a37c10cf67 Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.297165 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mwxh6"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.297933 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p4rz\" (UniqueName: \"kubernetes.io/projected/df7f22f3-54a7-4a2b-8ef9-fb22069e1f54-kube-api-access-5p4rz\") pod \"multus-admission-controller-857f4d67dd-np8cv\" (UID: \"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.299086 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.313283 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.317580 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sm96\" (UniqueName: \"kubernetes.io/projected/a9fd3151-6897-49e6-9c92-36b7dcb2953d-kube-api-access-2sm96\") pod \"ingress-canary-v8dh8\" (UID: \"a9fd3151-6897-49e6-9c92-36b7dcb2953d\") " pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.333210 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.361503 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.362168 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sns44\" (UniqueName: \"kubernetes.io/projected/2becf916-8ac8-48ad-ba1c-e679036dd069-kube-api-access-sns44\") pod \"control-plane-machine-set-operator-78cbb6b69f-c4z7l\" (UID: \"2becf916-8ac8-48ad-ba1c-e679036dd069\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.363327 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.369315 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.376611 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.377010 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.876981327 +0000 UTC m=+143.892111453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.377091 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.377690 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.877678158 +0000 UTC m=+143.892808284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.378293 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv9h8\" (UniqueName: \"kubernetes.io/projected/5cae4bdd-6891-40b0-9cf4-ba7cb676944f-kube-api-access-xv9h8\") pod \"service-ca-operator-777779d784-fdmsv\" (UID: \"5cae4bdd-6891-40b0-9cf4-ba7cb676944f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.383115 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zkrd\" (UniqueName: \"kubernetes.io/projected/b6b4beb5-c112-470d-9041-85f5ebd7cc42-kube-api-access-8zkrd\") pod \"csi-hostpathplugin-m8gjn\" (UID: \"b6b4beb5-c112-470d-9041-85f5ebd7cc42\") " pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.385584 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.389654 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjp7t\" (UniqueName: \"kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t\") pod \"marketplace-operator-79b997595-8k8fn\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.402758 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.412928 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldfzh\" (UniqueName: \"kubernetes.io/projected/ee9f4583-0a73-4090-a3e6-c0dada3788a6-kube-api-access-ldfzh\") pod \"package-server-manager-789f6589d5-j4tp6\" (UID: \"ee9f4583-0a73-4090-a3e6-c0dada3788a6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.426286 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48tdw\" (UniqueName: \"kubernetes.io/projected/498b1a02-6998-4001-962d-15a59163f760-kube-api-access-48tdw\") pod \"service-ca-9c57cc56f-lbcsz\" (UID: \"498b1a02-6998-4001-962d-15a59163f760\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.428235 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.434639 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaafca3f_0ac4_41c9_bc02_66fefa875aea.slice/crio-8848e1a8bb0e0b94fa7ac276e93499a9af28635671f8d9e5f92daa8515484a5c WatchSource:0}: Error finding container 8848e1a8bb0e0b94fa7ac276e93499a9af28635671f8d9e5f92daa8515484a5c: Status 404 returned error can't find the container with id 8848e1a8bb0e0b94fa7ac276e93499a9af28635671f8d9e5f92daa8515484a5c Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.435106 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe7cffc_33b6_4994_b2d0_1ae2294fc4ea.slice/crio-c1c90537d8f4334bb74e56945af77b32f7131b38b0916b42d9466b996a9a3f97 WatchSource:0}: Error finding container c1c90537d8f4334bb74e56945af77b32f7131b38b0916b42d9466b996a9a3f97: Status 404 returned error can't find the container with id c1c90537d8f4334bb74e56945af77b32f7131b38b0916b42d9466b996a9a3f97 Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.436850 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.443804 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.447473 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" event={"ID":"3ef5ad33-4ec5-4674-a517-8ade25ce2569","Type":"ContainerStarted","Data":"56858d58a43ab182f3f5491cb8259b7656ad7152a636b862b657b7d6bf1077cf"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.453624 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.460821 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zxbc\" (UniqueName: \"kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc\") pod \"collect-profiles-29494800-mpzgb\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.464006 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.468114 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" event={"ID":"a87d6206-4452-4dbc-a288-85c2a7b6332b","Type":"ContainerStarted","Data":"3d4e97ee8e942fe051d4399e9dd13535f6ff7ff1d48af72ec6810f6592cd1829"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.468149 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" event={"ID":"a87d6206-4452-4dbc-a288-85c2a7b6332b","Type":"ContainerStarted","Data":"c3a78d69ff8b1772be10b24eb8be18fdcf9509206323c2a5d497efb2aae08f42"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.470379 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.476614 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wclwf\" (UniqueName: \"kubernetes.io/projected/749b8957-29e5-434e-b736-3fd60318cb49-kube-api-access-wclwf\") pod \"machine-config-server-bfzr2\" (UID: \"749b8957-29e5-434e-b736-3fd60318cb49\") " pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.477794 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.478372 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.978328637 +0000 UTC m=+143.993458763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.478501 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.479113 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.479495 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:39.97948727 +0000 UTC m=+143.994617396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.487607 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.492896 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9zk6\" (UniqueName: \"kubernetes.io/projected/bc29514c-0991-488e-be96-46a03c78f543-kube-api-access-t9zk6\") pod \"dns-default-79rwf\" (UID: \"bc29514c-0991-488e-be96-46a03c78f543\") " pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.493975 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mwxh6" event={"ID":"b1ec8778-2a97-490f-a822-15d8bd226643","Type":"ContainerStarted","Data":"d5653e666226b402354b2375de332a6cd8de6ab6d552b32ac1f31429f3d4624c"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.495992 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v8dh8" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.499981 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pbdv4" event={"ID":"4c84abb7-3adc-4707-8aae-c2e8add2ad05","Type":"ContainerStarted","Data":"7c1e1b30d81f318073020956484aa23b562d8c78d1f07cdea8e85dd5531cb049"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.501493 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" event={"ID":"a7656713-548a-4e6b-aba3-2ca0f87d942f","Type":"ContainerStarted","Data":"ba649047098457db331f18b5afeee696176f0919bdfd2b5859faa3ae40017c76"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.501516 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" event={"ID":"a7656713-548a-4e6b-aba3-2ca0f87d942f","Type":"ContainerStarted","Data":"b7b89de3263111e39205cd5c46ef676f77db8de4ea49438423455ad555435b01"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.503033 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" event={"ID":"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb","Type":"ContainerStarted","Data":"68e3b0d2c7b5a8bae587797f20786e404c6a71d9f357b2052068f0a37c10cf67"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.503938 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" event={"ID":"7b20eefe-0735-4da1-9f09-0e455686b21e","Type":"ContainerStarted","Data":"0612b34ae46c1b18bcb47066160254fab8f7b0fe93d7841c5010e3f95256fd8b"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.505207 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s4sgr" event={"ID":"0517259c-9701-45de-9b5e-a71fb348fc2b","Type":"ContainerStarted","Data":"350d7ff5da6b6cc5508230b96726e6f75743ea1d3d072e7a5e47c2ad30506465"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.507484 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g7s6\" (UniqueName: \"kubernetes.io/projected/c67ab6d8-23c2-49de-b1b1-52a253777d13-kube-api-access-5g7s6\") pod \"olm-operator-6b444d44fb-qspck\" (UID: \"c67ab6d8-23c2-49de-b1b1-52a253777d13\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.508206 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-klxb6" event={"ID":"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4","Type":"ContainerStarted","Data":"7ad35825266d59a3862893709fb3a620f230213ff16f6ef75bb9f168b018bd54"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.510286 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" event={"ID":"b7559669-0e20-491a-95ec-3beacab1c1e1","Type":"ContainerStarted","Data":"7bc9e25c72d2ec3b3197ca3b756a667d21b97217ad8ba2d1de0c0aeaf2bb176d"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.510317 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" event={"ID":"b7559669-0e20-491a-95ec-3beacab1c1e1","Type":"ContainerStarted","Data":"4685306264dbc9f62ae1e7d3c33448c3fecbc4bde2fa144b4ec2db8e80b641ab"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.515057 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" event={"ID":"4598a597-b97b-4461-a1ea-05c049d93ebe","Type":"ContainerStarted","Data":"5ffcdec52360fa9ac1ff1156cdcc7830be1d6bfa39788f9e812fd52ed9a8c2bf"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.516492 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.519304 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" event={"ID":"d62f8973-63e6-4941-aa65-442c43e23b02","Type":"ContainerStarted","Data":"cc5ce3d2b61d346f212fe3fdde511a1df9d9e683f759f3946c6944b238abb8b0"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.521227 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" event={"ID":"887d412a-3e41-473f-9954-4fcfccf6b2ea","Type":"ContainerStarted","Data":"1e6729faa1ae6b24e4f34106769ab1384e011f7e7d99fd214720307aeff6e84c"} Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.523793 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bfzr2" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.580705 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.581232 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.081215431 +0000 UTC m=+144.096345557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.607901 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.683146 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.683789 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.183775886 +0000 UTC m=+144.198906012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.693225 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.711419 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kc9dp"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.711470 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.721051 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.724167 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x7jkw"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.744023 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc"] Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.752506 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda30945c5_89ca_4a8c_8568_863454de6d1f.slice/crio-03466edd99e36f6f909f3c36f1bd8ea7aae47a53c93e1c324eaa5221d305ef87 WatchSource:0}: Error finding container 03466edd99e36f6f909f3c36f1bd8ea7aae47a53c93e1c324eaa5221d305ef87: Status 404 returned error can't find the container with id 03466edd99e36f6f909f3c36f1bd8ea7aae47a53c93e1c324eaa5221d305ef87 Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.762044 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bn8qt"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.767629 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.784549 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.784961 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.28492416 +0000 UTC m=+144.300054296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.785380 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.785866 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.285809035 +0000 UTC m=+144.300939161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.798344 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j"] Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.841322 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38b7a4c3_bead_47c0_a37a_b0bf359b20aa.slice/crio-a118012ae11407ca58d44af9a9f6a3339cfa71072b09377720cdea961b48cc5f WatchSource:0}: Error finding container a118012ae11407ca58d44af9a9f6a3339cfa71072b09377720cdea961b48cc5f: Status 404 returned error can't find the container with id a118012ae11407ca58d44af9a9f6a3339cfa71072b09377720cdea961b48cc5f Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.869291 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfff9aa84_5f59_46e3_a40b_4622a6031d58.slice/crio-dadeb7776e666d7111c6ab25eb07292154fa71a30a4113da8517b5f3cf45c9f6 WatchSource:0}: Error finding container dadeb7776e666d7111c6ab25eb07292154fa71a30a4113da8517b5f3cf45c9f6: Status 404 returned error can't find the container with id dadeb7776e666d7111c6ab25eb07292154fa71a30a4113da8517b5f3cf45c9f6 Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.870415 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52adf4fb_76ea_4ca3_ab23_80c486b2a555.slice/crio-5b1b38064ac7caac9fcb0891e59375b55aedf0beb8dedc43fcea0f6467b97ca7 WatchSource:0}: Error finding container 5b1b38064ac7caac9fcb0891e59375b55aedf0beb8dedc43fcea0f6467b97ca7: Status 404 returned error can't find the container with id 5b1b38064ac7caac9fcb0891e59375b55aedf0beb8dedc43fcea0f6467b97ca7 Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.886495 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.886624 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.386595909 +0000 UTC m=+144.401726035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.888225 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.888582 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.388568066 +0000 UTC m=+144.403698192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.911528 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.928559 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx"] Jan 29 12:07:39 crc kubenswrapper[4993]: W0129 12:07:39.959047 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42b2535d_da68_4605_9c49_b8a3d530c094.slice/crio-b1a9d47e3155fcc45b59da1e6e91aa7cc92fd09cda0d2cead79c2ea3dbf8339f WatchSource:0}: Error finding container b1a9d47e3155fcc45b59da1e6e91aa7cc92fd09cda0d2cead79c2ea3dbf8339f: Status 404 returned error can't find the container with id b1a9d47e3155fcc45b59da1e6e91aa7cc92fd09cda0d2cead79c2ea3dbf8339f Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.983938 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-np8cv"] Jan 29 12:07:39 crc kubenswrapper[4993]: I0129 12:07:39.991472 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:39 crc kubenswrapper[4993]: E0129 12:07:39.992146 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.49212527 +0000 UTC m=+144.507255396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.037785 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.039053 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-llk59"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.073126 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.093170 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.093559 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.593537951 +0000 UTC m=+144.608668077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.130608 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7253cd01_1ade_43ba_a122_e2ba0f094648.slice/crio-003ee59ae0ef1d8dbb70f83f4170cf27e9fac42ddb9618dd46fa08d99069cbef WatchSource:0}: Error finding container 003ee59ae0ef1d8dbb70f83f4170cf27e9fac42ddb9618dd46fa08d99069cbef: Status 404 returned error can't find the container with id 003ee59ae0ef1d8dbb70f83f4170cf27e9fac42ddb9618dd46fa08d99069cbef Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.196769 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.197032 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.696989692 +0000 UTC m=+144.712119818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.197243 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.197965 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.6979572 +0000 UTC m=+144.713087316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.272838 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2becf916_8ac8_48ad_ba1c_e679036dd069.slice/crio-29cf9f2b34fbaeda10b4d2f9666c687477407d2e5e554bf0a3f375c256d6e14b WatchSource:0}: Error finding container 29cf9f2b34fbaeda10b4d2f9666c687477407d2e5e554bf0a3f375c256d6e14b: Status 404 returned error can't find the container with id 29cf9f2b34fbaeda10b4d2f9666c687477407d2e5e554bf0a3f375c256d6e14b Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.281937 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d86802e_e37a_42e9_aa5e_e0aff76ba6d7.slice/crio-3107fc8a8fa8da46cef44349a134712d484f5445cb149840c2c03a746026a9a5 WatchSource:0}: Error finding container 3107fc8a8fa8da46cef44349a134712d484f5445cb149840c2c03a746026a9a5: Status 404 returned error can't find the container with id 3107fc8a8fa8da46cef44349a134712d484f5445cb149840c2c03a746026a9a5 Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.301394 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.302212 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.802165882 +0000 UTC m=+144.817296008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.413247 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.413695 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:40.913680166 +0000 UTC m=+144.928810292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.424054 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8gjn"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.478448 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.493774 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.500374 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.505273 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v8dh8"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.513981 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.514126 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.014100228 +0000 UTC m=+145.029230354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.514553 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.514954 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.014945303 +0000 UTC m=+145.030075429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.560526 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" event={"ID":"7253cd01-1ade-43ba-a122-e2ba0f094648","Type":"ContainerStarted","Data":"003ee59ae0ef1d8dbb70f83f4170cf27e9fac42ddb9618dd46fa08d99069cbef"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.562216 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" event={"ID":"4598a597-b97b-4461-a1ea-05c049d93ebe","Type":"ContainerStarted","Data":"0cb2cee7bf1b114595f2e3ffdeab47ba8ba7472f12d079e8f61d945f1dd9b85a"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.563247 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bfzr2" event={"ID":"749b8957-29e5-434e-b736-3fd60318cb49","Type":"ContainerStarted","Data":"d8e3ecb7a7e2b0466466d2331ffb0365cef5993bc4d4c5df2a7b004bf614aa37"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.564062 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" event={"ID":"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea","Type":"ContainerStarted","Data":"c1c90537d8f4334bb74e56945af77b32f7131b38b0916b42d9466b996a9a3f97"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.564866 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" event={"ID":"eaafca3f-0ac4-41c9-bc02-66fefa875aea","Type":"ContainerStarted","Data":"8848e1a8bb0e0b94fa7ac276e93499a9af28635671f8d9e5f92daa8515484a5c"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.566305 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" event={"ID":"d62f8973-63e6-4941-aa65-442c43e23b02","Type":"ContainerStarted","Data":"add21d2a94271a16ec18c6721c654ee230f731dbf165d327132696d337ec7961"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.567115 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" event={"ID":"42b2535d-da68-4605-9c49-b8a3d530c094","Type":"ContainerStarted","Data":"b1a9d47e3155fcc45b59da1e6e91aa7cc92fd09cda0d2cead79c2ea3dbf8339f"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.567851 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" event={"ID":"38b7a4c3-bead-47c0-a37a-b0bf359b20aa","Type":"ContainerStarted","Data":"a118012ae11407ca58d44af9a9f6a3339cfa71072b09377720cdea961b48cc5f"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.569226 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" event={"ID":"ea098b70-73ee-4ee9-9f09-aa42910e215c","Type":"ContainerStarted","Data":"d53ff2b41f814d4e4167ca13485f5d7f0ca5a5cbbfe96502c0c747deaef1d4b7"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.570031 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" event={"ID":"0c2d1054-4b74-4702-b696-7cf26d9ff465","Type":"ContainerStarted","Data":"12a68ff7241bf72c1bf7ec919c77a187d8845b432d2d37b3ed91f898f4532e86"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.572343 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" event={"ID":"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7","Type":"ContainerStarted","Data":"3107fc8a8fa8da46cef44349a134712d484f5445cb149840c2c03a746026a9a5"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.574415 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" event={"ID":"a30945c5-89ca-4a8c-8568-863454de6d1f","Type":"ContainerStarted","Data":"03466edd99e36f6f909f3c36f1bd8ea7aae47a53c93e1c324eaa5221d305ef87"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.577286 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s4sgr" event={"ID":"0517259c-9701-45de-9b5e-a71fb348fc2b","Type":"ContainerStarted","Data":"86e61e722845488d2115bca21cb545f03521c2085112d9c7e637397f31ca7750"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.577968 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" event={"ID":"2becf916-8ac8-48ad-ba1c-e679036dd069","Type":"ContainerStarted","Data":"29cf9f2b34fbaeda10b4d2f9666c687477407d2e5e554bf0a3f375c256d6e14b"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.578624 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" event={"ID":"52adf4fb-76ea-4ca3-ab23-80c486b2a555","Type":"ContainerStarted","Data":"5b1b38064ac7caac9fcb0891e59375b55aedf0beb8dedc43fcea0f6467b97ca7"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.579448 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mwxh6" event={"ID":"b1ec8778-2a97-490f-a822-15d8bd226643","Type":"ContainerStarted","Data":"8586dd0994923ce240d58cde96aae60a13cc931eba36555d5b90f28935f75744"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.580213 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.581623 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.581671 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.585775 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" event={"ID":"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54","Type":"ContainerStarted","Data":"4ba520f4aafc157501cf81331513d941afc6c07a0d911b6abef1a0ab6576d669"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.589295 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" event={"ID":"887d412a-3e41-473f-9954-4fcfccf6b2ea","Type":"ContainerStarted","Data":"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.589500 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.590596 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" event={"ID":"5cae4bdd-6891-40b0-9cf4-ba7cb676944f","Type":"ContainerStarted","Data":"b776318c8caad7439a577dc37bd71ceae91fa116bb3335198753a34fabb7ff68"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.590880 4993 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9zsmx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.590915 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.591980 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" event={"ID":"fd0e07ac-1b26-4c71-9b89-9a19e45397ca","Type":"ContainerStarted","Data":"90b45edefb6f649c70d19ccc8a8c6e3219cb7a4c07ac627a40f4a19bcd0e23a8"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.594487 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" event={"ID":"fff9aa84-5f59-46e3-a40b-4622a6031d58","Type":"ContainerStarted","Data":"dadeb7776e666d7111c6ab25eb07292154fa71a30a4113da8517b5f3cf45c9f6"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.595759 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" event={"ID":"a0a5d6a8-59ec-4233-ae13-0404ea70c95c","Type":"ContainerStarted","Data":"cc47e4ea91406b610892aa0f88bf37f94e5385f7e7a1a4c63e8ac8ded215a086"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.608305 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" event={"ID":"7b20eefe-0735-4da1-9f09-0e455686b21e","Type":"ContainerStarted","Data":"7a2b73dbdcad02da398d52ceccd7ba79869d9cfad9606887913a9b8106ffa0ae"} Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.608452 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6b4beb5_c112_470d_9041_85f5ebd7cc42.slice/crio-03826d99463e72bb483cd7f32ba90575d5e048a29548225f0ff5ced63bc6445e WatchSource:0}: Error finding container 03826d99463e72bb483cd7f32ba90575d5e048a29548225f0ff5ced63bc6445e: Status 404 returned error can't find the container with id 03826d99463e72bb483cd7f32ba90575d5e048a29548225f0ff5ced63bc6445e Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.609402 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8835a1b6_a97d_48c8_9dd3_c197250e4825.slice/crio-99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5 WatchSource:0}: Error finding container 99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5: Status 404 returned error can't find the container with id 99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5 Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.609812 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" event={"ID":"0aa2a004-0aa7-499e-b2d6-6049dfbd9ddb","Type":"ContainerStarted","Data":"6b6d7c2d80e6405ff9a779ff2aa45c599f4ac345f05979079e063b1f61fda670"} Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.610628 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41f046c9_8758_4f06_a7d4_8149883734eb.slice/crio-6e83607e406fdb0d871560f73589ef1ccc8ce991b1d26432a497c0f50a84d134 WatchSource:0}: Error finding container 6e83607e406fdb0d871560f73589ef1ccc8ce991b1d26432a497c0f50a84d134: Status 404 returned error can't find the container with id 6e83607e406fdb0d871560f73589ef1ccc8ce991b1d26432a497c0f50a84d134 Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.612783 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pbdv4" event={"ID":"4c84abb7-3adc-4707-8aae-c2e8add2ad05","Type":"ContainerStarted","Data":"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.614903 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.616668 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.116653584 +0000 UTC m=+145.131783710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.624534 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb8104b0_8d5e_42e4_8912_859771cba6da.slice/crio-dc7738d8d9c4fbb9e1162c981f99378c33e666a48d012931142d8d4e112b81e5 WatchSource:0}: Error finding container dc7738d8d9c4fbb9e1162c981f99378c33e666a48d012931142d8d4e112b81e5: Status 404 returned error can't find the container with id dc7738d8d9c4fbb9e1162c981f99378c33e666a48d012931142d8d4e112b81e5 Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.624677 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-klxb6" event={"ID":"fd864d61-94e3-4ccf-b5cd-d0b42ca329a4","Type":"ContainerStarted","Data":"0925273e60c0af13066144fb6fb2e1d7c952d0ef341bab0eb6299eca10a49f0c"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.624889 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.626173 4993 patch_prober.go:28] interesting pod/console-operator-58897d9998-klxb6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.626237 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-klxb6" podUID="fd864d61-94e3-4ccf-b5cd-d0b42ca329a4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.629144 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" event={"ID":"3ef5ad33-4ec5-4674-a517-8ade25ce2569","Type":"ContainerStarted","Data":"c5709f04d49cbd29800122ddbed143d191211c44f8e0bfa4c5fa50c83c4d830a"} Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.629662 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.630925 4993 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kqng8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.630979 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.720690 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.723784 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.22376895 +0000 UTC m=+145.238899076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.728733 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.730421 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.751246 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbcsz"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.825798 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.827908 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-79rwf"] Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.832588 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.332560015 +0000 UTC m=+145.347690141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.832810 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.833130 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.333122641 +0000 UTC m=+145.348252767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.834510 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66"] Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.841522 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck"] Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.896495 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc29514c_0991_488e_be96_46a03c78f543.slice/crio-1bcd079c081a30e77e3028da2391f3966d3e24f96c4f2e027da70d11c1bf0b6f WatchSource:0}: Error finding container 1bcd079c081a30e77e3028da2391f3966d3e24f96c4f2e027da70d11c1bf0b6f: Status 404 returned error can't find the container with id 1bcd079c081a30e77e3028da2391f3966d3e24f96c4f2e027da70d11c1bf0b6f Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.900751 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9f4583_0a73_4090_a3e6_c0dada3788a6.slice/crio-abac55b50b3f368b3a61950b5538ae826095613250992b5ddb00acb4d54276cc WatchSource:0}: Error finding container abac55b50b3f368b3a61950b5538ae826095613250992b5ddb00acb4d54276cc: Status 404 returned error can't find the container with id abac55b50b3f368b3a61950b5538ae826095613250992b5ddb00acb4d54276cc Jan 29 12:07:40 crc kubenswrapper[4993]: W0129 12:07:40.904603 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacc7a110_91b0_4662_bdd2_0f7c998bc2e4.slice/crio-e1077ea9ac3779758ac24b4b9ef077ef88d20262a621f243c61426b5c288d39b WatchSource:0}: Error finding container e1077ea9ac3779758ac24b4b9ef077ef88d20262a621f243c61426b5c288d39b: Status 404 returned error can't find the container with id e1077ea9ac3779758ac24b4b9ef077ef88d20262a621f243c61426b5c288d39b Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.933935 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.934087 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.434059798 +0000 UTC m=+145.449189924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:40 crc kubenswrapper[4993]: I0129 12:07:40.935288 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:40 crc kubenswrapper[4993]: E0129 12:07:40.935705 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.435696066 +0000 UTC m=+145.450826242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.010641 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.012208 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.012255 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.025965 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" podStartSLOduration=124.025947815 podStartE2EDuration="2m4.025947815s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.023919796 +0000 UTC m=+145.039049922" watchObservedRunningTime="2026-01-29 12:07:41.025947815 +0000 UTC m=+145.041077941" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.036020 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.036174 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.53615277 +0000 UTC m=+145.551282906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.036346 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.036649 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.536640054 +0000 UTC m=+145.551770180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.068410 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-klxb6" podStartSLOduration=124.068392812 podStartE2EDuration="2m4.068392812s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.066904489 +0000 UTC m=+145.082034615" watchObservedRunningTime="2026-01-29 12:07:41.068392812 +0000 UTC m=+145.083522938" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.137891 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.138017 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.637989644 +0000 UTC m=+145.653119770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.138264 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.138606 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.638594531 +0000 UTC m=+145.653724657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.142418 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mwxh6" podStartSLOduration=124.142391651 podStartE2EDuration="2m4.142391651s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.140824456 +0000 UTC m=+145.155954582" watchObservedRunningTime="2026-01-29 12:07:41.142391651 +0000 UTC m=+145.157521777" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.183016 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-c8dv2" podStartSLOduration=124.182999995 podStartE2EDuration="2m4.182999995s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.181598404 +0000 UTC m=+145.196728520" watchObservedRunningTime="2026-01-29 12:07:41.182999995 +0000 UTC m=+145.198130121" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.225359 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podStartSLOduration=123.225337739 podStartE2EDuration="2m3.225337739s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.221665253 +0000 UTC m=+145.236795389" watchObservedRunningTime="2026-01-29 12:07:41.225337739 +0000 UTC m=+145.240467865" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.240209 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.240689 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.740671832 +0000 UTC m=+145.755801958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.321503 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-h88fc" podStartSLOduration=124.321482868 podStartE2EDuration="2m4.321482868s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.320290524 +0000 UTC m=+145.335420670" watchObservedRunningTime="2026-01-29 12:07:41.321482868 +0000 UTC m=+145.336612994" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.341815 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.342213 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.842197747 +0000 UTC m=+145.857327873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.404416 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzcct" podStartSLOduration=124.404393675 podStartE2EDuration="2m4.404393675s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.400545143 +0000 UTC m=+145.415675289" watchObservedRunningTime="2026-01-29 12:07:41.404393675 +0000 UTC m=+145.419523801" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.442717 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.443198 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.943141255 +0000 UTC m=+145.958271381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.443430 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.443775 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:41.943762113 +0000 UTC m=+145.958892239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.531318 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-s4sgr" podStartSLOduration=124.531298774 podStartE2EDuration="2m4.531298774s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.526749322 +0000 UTC m=+145.541879468" watchObservedRunningTime="2026-01-29 12:07:41.531298774 +0000 UTC m=+145.546428900" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.545119 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.545593 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.045560466 +0000 UTC m=+146.060690602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.568059 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pbdv4" podStartSLOduration=124.568043706 podStartE2EDuration="2m4.568043706s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.564021499 +0000 UTC m=+145.579151625" watchObservedRunningTime="2026-01-29 12:07:41.568043706 +0000 UTC m=+145.583173832" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.635389 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" event={"ID":"a0a5d6a8-59ec-4233-ae13-0404ea70c95c","Type":"ContainerStarted","Data":"973d7a6eb6404e81899f92674f5a4ebafba485d146ae84a3c3dc5fb42a6cffc3"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.636502 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" event={"ID":"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7","Type":"ContainerStarted","Data":"1349e6f7b565e31dff59798d0838c56b3af4083ca5a8f3711b6190652f1ac42a"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.644357 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" event={"ID":"5cae4bdd-6891-40b0-9cf4-ba7cb676944f","Type":"ContainerStarted","Data":"4ea2ed530620a01e828f89e837b458465a096deccb0d120b662b6339f7be2948"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.645579 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" event={"ID":"8835a1b6-a97d-48c8-9dd3-c197250e4825","Type":"ContainerStarted","Data":"ccfc6e9565d9325d95e7125b923c9969c9c76b944792c576c6e12306ce990dd4"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.645623 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" event={"ID":"8835a1b6-a97d-48c8-9dd3-c197250e4825","Type":"ContainerStarted","Data":"99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.648489 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.649242 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.149222962 +0000 UTC m=+146.164353088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.651151 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" event={"ID":"fb8104b0-8d5e-42e4-8912-859771cba6da","Type":"ContainerStarted","Data":"dc7738d8d9c4fbb9e1162c981f99378c33e666a48d012931142d8d4e112b81e5"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.659086 4993 generic.go:334] "Generic (PLEG): container finished" podID="52adf4fb-76ea-4ca3-ab23-80c486b2a555" containerID="ddaceb91d58b514a8401d9e27019626687b15f98980b1890f097ff712ef809f8" exitCode=0 Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.659522 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" event={"ID":"52adf4fb-76ea-4ca3-ab23-80c486b2a555","Type":"ContainerDied","Data":"ddaceb91d58b514a8401d9e27019626687b15f98980b1890f097ff712ef809f8"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.661298 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" event={"ID":"fff9aa84-5f59-46e3-a40b-4622a6031d58","Type":"ContainerStarted","Data":"ddcf4b11dd71f6c287617e906890156072d77647f02736299f49a1bd4d016e6e"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.662578 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" event={"ID":"b6b4beb5-c112-470d-9041-85f5ebd7cc42","Type":"ContainerStarted","Data":"03826d99463e72bb483cd7f32ba90575d5e048a29548225f0ff5ced63bc6445e"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.664264 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" event={"ID":"5fe7cffc-33b6-4994-b2d0-1ae2294fc4ea","Type":"ContainerStarted","Data":"8e4c0cc64bd97924a5437d1320ab93a4ea239ccdfb82e4d29a10c70ce0388313"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.676808 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" event={"ID":"498b1a02-6998-4001-962d-15a59163f760","Type":"ContainerStarted","Data":"92df976812fa524494b782d625cbfcc396873347eee4b3ea1bd3356fcfbff4b3"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.680307 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v8dh8" event={"ID":"a9fd3151-6897-49e6-9c92-36b7dcb2953d","Type":"ContainerStarted","Data":"f1f7dc162cd4c6c324509fe5586b7f9457f86924017df43f3b75ed10e13cb5bc"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.686095 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" event={"ID":"fd0e07ac-1b26-4c71-9b89-9a19e45397ca","Type":"ContainerStarted","Data":"86210fa43b6d8edd62b6cca356d1c4885798baf2548ac825557846406eb79c20"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.687771 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79rwf" event={"ID":"bc29514c-0991-488e-be96-46a03c78f543","Type":"ContainerStarted","Data":"1bcd079c081a30e77e3028da2391f3966d3e24f96c4f2e027da70d11c1bf0b6f"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.689400 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" event={"ID":"38b7a4c3-bead-47c0-a37a-b0bf359b20aa","Type":"ContainerStarted","Data":"97038fe9f4931931b874fb086fd2c9ff4328e83574b89efa1064fd7787e3f68e"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.690585 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" event={"ID":"eaafca3f-0ac4-41c9-bc02-66fefa875aea","Type":"ContainerStarted","Data":"e80c88422fca2095de9b18876e043d3e0997ff9129a4a83f32a504353d2b3e8a"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.691332 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" event={"ID":"9f938607-0656-46f5-860b-45e736c42c1b","Type":"ContainerStarted","Data":"a002478e7c6cce5ed3a2943eefd0520688ce630ca96e449782564c34a754705b"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.692568 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" event={"ID":"42b2535d-da68-4605-9c49-b8a3d530c094","Type":"ContainerStarted","Data":"880fc2e13f2ab1ceccd7042b933dc02b5d2a5f2f19ea7eb56fd04704cb6d540d"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.693815 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" event={"ID":"3ef5ad33-4ec5-4674-a517-8ade25ce2569","Type":"ContainerStarted","Data":"a5ed91c51dc04e859e32090a50bcc313e41a5421f520a15c2cb433c2840886b0"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.694676 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bfzr2" event={"ID":"749b8957-29e5-434e-b736-3fd60318cb49","Type":"ContainerStarted","Data":"36e24f5c152d92903ed411b0e7ed2b78781a76334c23bf64c2046c25d39845a7"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.695572 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" event={"ID":"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54","Type":"ContainerStarted","Data":"398eee9646b0544f2468d518843f8af3667bb0d7e9f3189e5e5f052bee671589"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.696215 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" event={"ID":"acc7a110-91b0-4662-bdd2-0f7c998bc2e4","Type":"ContainerStarted","Data":"e1077ea9ac3779758ac24b4b9ef077ef88d20262a621f243c61426b5c288d39b"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.696938 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" event={"ID":"0c2d1054-4b74-4702-b696-7cf26d9ff465","Type":"ContainerStarted","Data":"0f26dd972aa3c1766c522e74a799b1012eb055cd8b65ddfffccded9bc308103d"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.697502 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" event={"ID":"c67ab6d8-23c2-49de-b1b1-52a253777d13","Type":"ContainerStarted","Data":"64de69dc107a3faa18ae90b2bc7ea62601375eca708193ef71f5078e5c60ce09"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.698198 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" event={"ID":"2becf916-8ac8-48ad-ba1c-e679036dd069","Type":"ContainerStarted","Data":"c37e0373377e52895452556c374c97349cf1499c3b94ba8b649e6afed86eb8d5"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.698938 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" event={"ID":"a30945c5-89ca-4a8c-8568-863454de6d1f","Type":"ContainerStarted","Data":"ba58838f711df1e8fbf77f03879b0e11fa67a87b550559c4c3953366ab3b63bd"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.699851 4993 generic.go:334] "Generic (PLEG): container finished" podID="7b20eefe-0735-4da1-9f09-0e455686b21e" containerID="7a2b73dbdcad02da398d52ceccd7ba79869d9cfad9606887913a9b8106ffa0ae" exitCode=0 Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.699898 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" event={"ID":"7b20eefe-0735-4da1-9f09-0e455686b21e","Type":"ContainerDied","Data":"7a2b73dbdcad02da398d52ceccd7ba79869d9cfad9606887913a9b8106ffa0ae"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.701752 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" event={"ID":"a7656713-548a-4e6b-aba3-2ca0f87d942f","Type":"ContainerStarted","Data":"83fa6b3e0685acbff1ee9d3c357d8d87d72b347b5b8ee6c23f4b02ea46fad0db"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.703138 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" event={"ID":"ea098b70-73ee-4ee9-9f09-aa42910e215c","Type":"ContainerStarted","Data":"083a9152185eb737c3ad127fd254e8b9b5c8c874168a08f58c389c25ab9d2d51"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.704022 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerStarted","Data":"6e83607e406fdb0d871560f73589ef1ccc8ce991b1d26432a497c0f50a84d134"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705027 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" event={"ID":"ee9f4583-0a73-4090-a3e6-c0dada3788a6","Type":"ContainerStarted","Data":"abac55b50b3f368b3a61950b5538ae826095613250992b5ddb00acb4d54276cc"} Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705799 4993 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9zsmx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705836 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705799 4993 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kqng8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705878 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705799 4993 patch_prober.go:28] interesting pod/console-operator-58897d9998-klxb6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.705905 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-klxb6" podUID="fd864d61-94e3-4ccf-b5cd-d0b42ca329a4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.706491 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.707464 4993 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-md8vs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.707787 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.708030 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.708084 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.749495 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.749615 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.249597284 +0000 UTC m=+146.264727410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.750297 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.754844 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.254822074 +0000 UTC m=+146.269952200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.845041 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-q2w6x" podStartSLOduration=124.845023283 podStartE2EDuration="2m4.845023283s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.844500968 +0000 UTC m=+145.859631094" watchObservedRunningTime="2026-01-29 12:07:41.845023283 +0000 UTC m=+145.860153409" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.853303 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.853649 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.353635121 +0000 UTC m=+146.368765247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.884809 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rzb4l" podStartSLOduration=124.884795602 podStartE2EDuration="2m4.884795602s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.881790255 +0000 UTC m=+145.896920381" watchObservedRunningTime="2026-01-29 12:07:41.884795602 +0000 UTC m=+145.899925728" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.925554 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podStartSLOduration=124.92553957 podStartE2EDuration="2m4.92553957s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:41.922834601 +0000 UTC m=+145.937964727" watchObservedRunningTime="2026-01-29 12:07:41.92553957 +0000 UTC m=+145.940669696" Jan 29 12:07:41 crc kubenswrapper[4993]: I0129 12:07:41.958779 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:41 crc kubenswrapper[4993]: E0129 12:07:41.959100 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.459087859 +0000 UTC m=+146.474217985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.006412 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4bnwv" podStartSLOduration=125.006393287 podStartE2EDuration="2m5.006393287s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.005643496 +0000 UTC m=+146.020773622" watchObservedRunningTime="2026-01-29 12:07:42.006393287 +0000 UTC m=+146.021523413" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.011028 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.011081 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.055714 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-x7jkw" podStartSLOduration=125.055695102 podStartE2EDuration="2m5.055695102s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.054930511 +0000 UTC m=+146.070060637" watchObservedRunningTime="2026-01-29 12:07:42.055695102 +0000 UTC m=+146.070825228" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.059747 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.060145 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.560129321 +0000 UTC m=+146.575259447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.085647 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2ztwc" podStartSLOduration=125.085629648 podStartE2EDuration="2m5.085629648s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.084363481 +0000 UTC m=+146.099493607" watchObservedRunningTime="2026-01-29 12:07:42.085629648 +0000 UTC m=+146.100759774" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.163870 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.164177 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.664166148 +0000 UTC m=+146.679296274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.264406 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.264616 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.764590562 +0000 UTC m=+146.779720698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.264938 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.265285 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.765275051 +0000 UTC m=+146.780405177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.366274 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.366703 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.866673741 +0000 UTC m=+146.881803867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.468039 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.468425 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:42.968409943 +0000 UTC m=+146.983540069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.569168 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.569575 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.069557366 +0000 UTC m=+147.084687492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.670577 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.670882 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.170866846 +0000 UTC m=+147.185996972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.710021 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" event={"ID":"498b1a02-6998-4001-962d-15a59163f760","Type":"ContainerStarted","Data":"fb8e0b5c17800fb0fa2dae8611b32908b8580390d01bdda9033641bd15315f33"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.711021 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v8dh8" event={"ID":"a9fd3151-6897-49e6-9c92-36b7dcb2953d","Type":"ContainerStarted","Data":"faff1deb6b04ac73cc2276baece055625af74dfa10f8e1680393e09ad5424715"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.712779 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" event={"ID":"7b20eefe-0735-4da1-9f09-0e455686b21e","Type":"ContainerStarted","Data":"ba18e5a6b9cd7fd6e2016c488a6cfb012e25249bf3fe6e192ba88e9f9275e4a4"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.714025 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" event={"ID":"a30945c5-89ca-4a8c-8568-863454de6d1f","Type":"ContainerStarted","Data":"6c48c42996686102c75e86f50a05d2e8c8b49b3fbbac5af9464b2b03b56fadc1"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.715468 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" event={"ID":"acc7a110-91b0-4662-bdd2-0f7c998bc2e4","Type":"ContainerStarted","Data":"823bc0a612b09327eaa9ccf9b238f0dcd0d6c9e3202322664f0594ad67a45179"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.716813 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" event={"ID":"7253cd01-1ade-43ba-a122-e2ba0f094648","Type":"ContainerStarted","Data":"dd5adcfaed8dcfee96f516919117afc041cd1853c8c97f0f4b2ab4d7b8cfefe5"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.719503 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79rwf" event={"ID":"bc29514c-0991-488e-be96-46a03c78f543","Type":"ContainerStarted","Data":"d3aeac8e9562a7b7b72446b09e3314b0b6ceef8b54e46c3390ef3e5b1616c12a"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.720883 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" event={"ID":"c67ab6d8-23c2-49de-b1b1-52a253777d13","Type":"ContainerStarted","Data":"9d966f4dc875d1dc5db4d21dc6b66797a29955dde5891642867dd24d7984a00d"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.722738 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" event={"ID":"0c2d1054-4b74-4702-b696-7cf26d9ff465","Type":"ContainerStarted","Data":"4a319410740aea648b58b830a18301224e358cdf2e8f6a61c4b54f6607570b4f"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.724447 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" event={"ID":"9f938607-0656-46f5-860b-45e736c42c1b","Type":"ContainerStarted","Data":"41abccda33d214697730d4500abb85678d93d63850886f8813dccac2b45b5911"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.725960 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" event={"ID":"ee9f4583-0a73-4090-a3e6-c0dada3788a6","Type":"ContainerStarted","Data":"cff28a1d47fe5a107a35d25c71b377a18178679ede5e8d478076f62b91f26170"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.727808 4993 generic.go:334] "Generic (PLEG): container finished" podID="38b7a4c3-bead-47c0-a37a-b0bf359b20aa" containerID="97038fe9f4931931b874fb086fd2c9ff4328e83574b89efa1064fd7787e3f68e" exitCode=0 Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.727910 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" event={"ID":"38b7a4c3-bead-47c0-a37a-b0bf359b20aa","Type":"ContainerDied","Data":"97038fe9f4931931b874fb086fd2c9ff4328e83574b89efa1064fd7787e3f68e"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.729465 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" event={"ID":"fb8104b0-8d5e-42e4-8912-859771cba6da","Type":"ContainerStarted","Data":"1f43999c547e78119ccfa373a6d3c3acaef08056f9eafc3512b25a62faa9f55b"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.730841 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerStarted","Data":"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904"} Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.731629 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.731661 4993 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-md8vs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.731676 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.731705 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.735353 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-v8dh8" podStartSLOduration=6.73534113 podStartE2EDuration="6.73534113s" podCreationTimestamp="2026-01-29 12:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.732539078 +0000 UTC m=+146.747669204" watchObservedRunningTime="2026-01-29 12:07:42.73534113 +0000 UTC m=+146.750471256" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.737175 4993 patch_prober.go:28] interesting pod/console-operator-58897d9998-klxb6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.737246 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-klxb6" podUID="fd864d61-94e3-4ccf-b5cd-d0b42ca329a4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.771751 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.773166 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.273142532 +0000 UTC m=+147.288272758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.792737 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fdmsv" podStartSLOduration=124.792720978 podStartE2EDuration="2m4.792720978s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.792523323 +0000 UTC m=+146.807653449" watchObservedRunningTime="2026-01-29 12:07:42.792720978 +0000 UTC m=+146.807851104" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.793121 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" podStartSLOduration=125.793117279 podStartE2EDuration="2m5.793117279s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.765968995 +0000 UTC m=+146.781099121" watchObservedRunningTime="2026-01-29 12:07:42.793117279 +0000 UTC m=+146.808247405" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.839083 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bfzr2" podStartSLOduration=6.839042227 podStartE2EDuration="6.839042227s" podCreationTimestamp="2026-01-29 12:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.819282526 +0000 UTC m=+146.834412652" watchObservedRunningTime="2026-01-29 12:07:42.839042227 +0000 UTC m=+146.854172353" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.873921 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.874322 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.374310107 +0000 UTC m=+147.389440233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.881570 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-47m4j" podStartSLOduration=125.881548506 podStartE2EDuration="2m5.881548506s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.840244822 +0000 UTC m=+146.855374948" watchObservedRunningTime="2026-01-29 12:07:42.881548506 +0000 UTC m=+146.896678632" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.928467 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xrmnd" podStartSLOduration=125.928442962 podStartE2EDuration="2m5.928442962s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.919124992 +0000 UTC m=+146.934255128" watchObservedRunningTime="2026-01-29 12:07:42.928442962 +0000 UTC m=+146.943573088" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.969451 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c4z7l" podStartSLOduration=125.969434376 podStartE2EDuration="2m5.969434376s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.966556033 +0000 UTC m=+146.981686159" watchObservedRunningTime="2026-01-29 12:07:42.969434376 +0000 UTC m=+146.984564502" Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.980338 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:42 crc kubenswrapper[4993]: E0129 12:07:42.983534 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.483518683 +0000 UTC m=+147.498648809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:42 crc kubenswrapper[4993]: I0129 12:07:42.996736 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" podStartSLOduration=125.996719665 podStartE2EDuration="2m5.996719665s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:42.996375415 +0000 UTC m=+147.011505551" watchObservedRunningTime="2026-01-29 12:07:42.996719665 +0000 UTC m=+147.011849791" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.014368 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.014436 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.023523 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wfg4q" podStartSLOduration=126.02350752 podStartE2EDuration="2m6.02350752s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.019574806 +0000 UTC m=+147.034704932" watchObservedRunningTime="2026-01-29 12:07:43.02350752 +0000 UTC m=+147.038637646" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.085523 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.085880 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.585868042 +0000 UTC m=+147.600998168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.187489 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.187689 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.687659595 +0000 UTC m=+147.702789721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.188040 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.188355 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.688342995 +0000 UTC m=+147.703473121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.289342 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.289482 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.789462878 +0000 UTC m=+147.804593014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.289507 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.289873 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.78986521 +0000 UTC m=+147.804995336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.390863 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.391049 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.891022494 +0000 UTC m=+147.906152630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.391250 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.391550 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.891538448 +0000 UTC m=+147.906668574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.492207 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.492419 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.992388104 +0000 UTC m=+148.007518230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.492500 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.493108 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:43.993098115 +0000 UTC m=+148.008228241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.594104 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.594317 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.09429317 +0000 UTC m=+148.109423296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.594481 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.594757 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.094743563 +0000 UTC m=+148.109873689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.695725 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.695955 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.195923388 +0000 UTC m=+148.211053524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.696363 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.696623 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.196611918 +0000 UTC m=+148.211742044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.736577 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" event={"ID":"9d86802e-e37a-42e9-aa5e-e0aff76ba6d7","Type":"ContainerStarted","Data":"baf2e0ccd0192a9a2a8d952b4682712259fb1912f32e41b3718b5bd4667c654a"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.739223 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" event={"ID":"ee9f4583-0a73-4090-a3e6-c0dada3788a6","Type":"ContainerStarted","Data":"62a948eb8828a624e3c89c543c831030237dca86c9f231fb574f21b0dd7875a5"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.741287 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" event={"ID":"fff9aa84-5f59-46e3-a40b-4622a6031d58","Type":"ContainerStarted","Data":"1bb91e7de781451188d11f8263301d5ddba2a6d05934788029507c0f9f261aae"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.743200 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" event={"ID":"38b7a4c3-bead-47c0-a37a-b0bf359b20aa","Type":"ContainerStarted","Data":"c7d9482cc363875b2b36f4f52b0fb1ad836d212459ea8b4bf4b5da63f90dff84"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.744674 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" event={"ID":"acc7a110-91b0-4662-bdd2-0f7c998bc2e4","Type":"ContainerStarted","Data":"871e08e8d7a7dbe489b927bbf2c755663d958df9367c15b37fd1b442a7b43100"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.746873 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" event={"ID":"ea098b70-73ee-4ee9-9f09-aa42910e215c","Type":"ContainerStarted","Data":"bd532ea7e23200faa63dd171f34e655ba664bfa6dd037441d482f6fd36c70d2f"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.749135 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" event={"ID":"52adf4fb-76ea-4ca3-ab23-80c486b2a555","Type":"ContainerStarted","Data":"2eafed1d1588303ef224a4485491b1010777d138c87748af579b27b825e40e69"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.753928 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" event={"ID":"df7f22f3-54a7-4a2b-8ef9-fb22069e1f54","Type":"ContainerStarted","Data":"19552d5c4ff7a6e04f420826561e71531a83738aca2894d20f8d146b2e7bb81f"} Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754008 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754022 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754014 4993 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7nkd9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754067 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" podUID="9f938607-0656-46f5-860b-45e736c42c1b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754370 4993 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8k8fn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.754432 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.759090 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-llk59" podStartSLOduration=126.759074724 podStartE2EDuration="2m6.759074724s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.7582634 +0000 UTC m=+147.773393546" watchObservedRunningTime="2026-01-29 12:07:43.759074724 +0000 UTC m=+147.774204850" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.797220 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.798488 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.298470692 +0000 UTC m=+148.313600828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.802290 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" podStartSLOduration=125.802272213 podStartE2EDuration="2m5.802272213s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.800904933 +0000 UTC m=+147.816035059" watchObservedRunningTime="2026-01-29 12:07:43.802272213 +0000 UTC m=+147.817402349" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.839066 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" podStartSLOduration=125.839051476 podStartE2EDuration="2m5.839051476s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.838064937 +0000 UTC m=+147.853195073" watchObservedRunningTime="2026-01-29 12:07:43.839051476 +0000 UTC m=+147.854181602" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.861551 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" podStartSLOduration=126.861530695 podStartE2EDuration="2m6.861530695s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.858353644 +0000 UTC m=+147.873483790" watchObservedRunningTime="2026-01-29 12:07:43.861530695 +0000 UTC m=+147.876660821" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.875823 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" podStartSLOduration=125.875806678 podStartE2EDuration="2m5.875806678s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.874969204 +0000 UTC m=+147.890099330" watchObservedRunningTime="2026-01-29 12:07:43.875806678 +0000 UTC m=+147.890936804" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.890715 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bn8qt" podStartSLOduration=126.890696309 podStartE2EDuration="2m6.890696309s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.889890115 +0000 UTC m=+147.905020231" watchObservedRunningTime="2026-01-29 12:07:43.890696309 +0000 UTC m=+147.905826435" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.899172 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:43 crc kubenswrapper[4993]: E0129 12:07:43.900598 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.400584674 +0000 UTC m=+148.415714800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.919119 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rxfxk" podStartSLOduration=126.919099389 podStartE2EDuration="2m6.919099389s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.912757257 +0000 UTC m=+147.927887383" watchObservedRunningTime="2026-01-29 12:07:43.919099389 +0000 UTC m=+147.934229515" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.938173 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d8nzx" podStartSLOduration=126.938154881 podStartE2EDuration="2m6.938154881s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.936531124 +0000 UTC m=+147.951661240" watchObservedRunningTime="2026-01-29 12:07:43.938154881 +0000 UTC m=+147.953285007" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.977653 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-lbcsz" podStartSLOduration=125.977639422 podStartE2EDuration="2m5.977639422s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.974023557 +0000 UTC m=+147.989153683" watchObservedRunningTime="2026-01-29 12:07:43.977639422 +0000 UTC m=+147.992769548" Jan 29 12:07:43 crc kubenswrapper[4993]: I0129 12:07:43.990748 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2fbch" podStartSLOduration=126.99072699 podStartE2EDuration="2m6.99072699s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:43.989549337 +0000 UTC m=+148.004679463" watchObservedRunningTime="2026-01-29 12:07:43.99072699 +0000 UTC m=+148.005857116" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:43.999910 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.000254 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.500239855 +0000 UTC m=+148.515369991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.007483 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9zb66" podStartSLOduration=127.007468014 podStartE2EDuration="2m7.007468014s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.006339382 +0000 UTC m=+148.021469508" watchObservedRunningTime="2026-01-29 12:07:44.007468014 +0000 UTC m=+148.022598150" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.014861 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:44 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:44 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:44 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.014913 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.026147 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-np8cv" podStartSLOduration=127.026130904 podStartE2EDuration="2m7.026130904s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.024802175 +0000 UTC m=+148.039932301" watchObservedRunningTime="2026-01-29 12:07:44.026130904 +0000 UTC m=+148.041261030" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.101983 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.102456 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.602438409 +0000 UTC m=+148.617568535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.203074 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.203339 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.703310945 +0000 UTC m=+148.718441071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.203566 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.203909 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.703897953 +0000 UTC m=+148.719028079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.305031 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.305228 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.805199381 +0000 UTC m=+148.820329507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.305286 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.305568 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.805555841 +0000 UTC m=+148.820685967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.406838 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.407057 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.907027755 +0000 UTC m=+148.922157911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.407197 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.407464 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:44.907451517 +0000 UTC m=+148.922581633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.508815 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.509026 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.008990762 +0000 UTC m=+149.024120898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.509130 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.509481 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.009471396 +0000 UTC m=+149.024601622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.601007 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.610725 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.610906 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.110885617 +0000 UTC m=+149.126015743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.611050 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.611362 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.111350661 +0000 UTC m=+149.126480787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.712443 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.712620 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.212594708 +0000 UTC m=+149.227724834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.712953 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.713286 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.213278667 +0000 UTC m=+149.228408793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.717629 4993 csr.go:261] certificate signing request csr-zcpth is approved, waiting to be issued Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.726687 4993 csr.go:257] certificate signing request csr-zcpth is issued Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.760861 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" event={"ID":"38b7a4c3-bead-47c0-a37a-b0bf359b20aa","Type":"ContainerStarted","Data":"9a15dcf22743b3fef6aa227bc99e345f5759b429954af724ed807cfd1191baa2"} Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.762212 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" event={"ID":"b6b4beb5-c112-470d-9041-85f5ebd7cc42","Type":"ContainerStarted","Data":"2fb072ad5cfed0f9fd0f90d022e56ab9a910baad1179e4e0f11dc3c1649c9133"} Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.763921 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79rwf" event={"ID":"bc29514c-0991-488e-be96-46a03c78f543","Type":"ContainerStarted","Data":"18a39856adc4f7a358355d68f68850dfcae7c66206e92d08322426ced8da989e"} Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.764927 4993 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8k8fn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.764966 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.765759 4993 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7nkd9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.765809 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" podUID="9f938607-0656-46f5-860b-45e736c42c1b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.813653 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.814037 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.314022499 +0000 UTC m=+149.329152615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.816652 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" podStartSLOduration=127.816640685 podStartE2EDuration="2m7.816640685s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.795209336 +0000 UTC m=+148.810339482" watchObservedRunningTime="2026-01-29 12:07:44.816640685 +0000 UTC m=+148.831770811" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.838370 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" podStartSLOduration=126.838355093 podStartE2EDuration="2m6.838355093s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.818421277 +0000 UTC m=+148.833551403" watchObservedRunningTime="2026-01-29 12:07:44.838355093 +0000 UTC m=+148.853485219" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.839341 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" podStartSLOduration=126.839336391 podStartE2EDuration="2m6.839336391s" podCreationTimestamp="2026-01-29 12:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.836789458 +0000 UTC m=+148.851919584" watchObservedRunningTime="2026-01-29 12:07:44.839336391 +0000 UTC m=+148.854466517" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.856213 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-79rwf" podStartSLOduration=8.856198219 podStartE2EDuration="8.856198219s" podCreationTimestamp="2026-01-29 12:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:44.855049116 +0000 UTC m=+148.870179242" watchObservedRunningTime="2026-01-29 12:07:44.856198219 +0000 UTC m=+148.871328345" Jan 29 12:07:44 crc kubenswrapper[4993]: I0129 12:07:44.915953 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:44 crc kubenswrapper[4993]: E0129 12:07:44.919469 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.419453838 +0000 UTC m=+149.434583964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.017309 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.017684 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.517668887 +0000 UTC m=+149.532799013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.019033 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:45 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:45 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:45 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.019072 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.119000 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.119358 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.619347575 +0000 UTC m=+149.634477701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.220163 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.220501 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.72048501 +0000 UTC m=+149.735615136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.321212 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.321507 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.82149488 +0000 UTC m=+149.836625006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.422486 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.422780 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:45.922765907 +0000 UTC m=+149.937896033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.523812 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.524216 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.02420151 +0000 UTC m=+150.039331636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.625319 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.625804 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.125784205 +0000 UTC m=+150.140914341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.693858 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.726686 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.727053 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.227036152 +0000 UTC m=+150.242166278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.728363 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-29 12:02:44 +0000 UTC, rotation deadline is 2026-12-18 11:57:58.633485384 +0000 UTC Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.728388 4993 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7751h50m12.905100683s for next certificate rotation Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.827989 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.828896 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.328880246 +0000 UTC m=+150.344010372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:45 crc kubenswrapper[4993]: I0129 12:07:45.929679 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:45 crc kubenswrapper[4993]: E0129 12:07:45.930088 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.430068861 +0000 UTC m=+150.445198997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.013305 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:46 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:46 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:46 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.013357 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.030624 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.031037 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.53101639 +0000 UTC m=+150.546146516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.131911 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.132347 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.632331329 +0000 UTC m=+150.647461455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.232828 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.233067 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.233110 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.233673 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.733656727 +0000 UTC m=+150.748786853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.234343 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.239891 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.334853 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.334947 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.334980 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.335609 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.835592734 +0000 UTC m=+150.850722860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.338503 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.342707 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.437047 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.437280 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.937247023 +0000 UTC m=+150.952377159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.437424 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.437879 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:46.937870081 +0000 UTC m=+150.953000207 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.506115 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.518198 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.522635 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.543757 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.543944 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.043918137 +0000 UTC m=+151.059048263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.544044 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.544514 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.044496723 +0000 UTC m=+151.059626849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.645538 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.645734 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.145698879 +0000 UTC m=+151.160829005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.645906 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.646174 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.146165962 +0000 UTC m=+151.161296088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.673064 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.673967 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.679285 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.698142 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.747263 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.747478 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.24743876 +0000 UTC m=+151.262568886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.747543 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt8jz\" (UniqueName: \"kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.747705 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.747728 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.747987 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.748296 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.248281764 +0000 UTC m=+151.263411890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.811047 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" event={"ID":"b6b4beb5-c112-470d-9041-85f5ebd7cc42","Type":"ContainerStarted","Data":"2c92f92b43d50559609e48868e8eed2758680e59706eebb60f01f30cb3d7e22a"} Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.848705 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.848919 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt8jz\" (UniqueName: \"kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.848971 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.848994 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.849508 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.849586 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.349569112 +0000 UTC m=+151.364699238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.850109 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.950311 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt8jz\" (UniqueName: \"kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz\") pod \"certified-operators-zzw5p\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.952131 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:46 crc kubenswrapper[4993]: E0129 12:07:46.952573 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.45255522 +0000 UTC m=+151.467685346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:46 crc kubenswrapper[4993]: I0129 12:07:46.993428 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.004692 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.005292 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.007600 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.023076 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:47 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:47 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:47 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.023123 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.023579 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.040801 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.054336 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.054588 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.554571219 +0000 UTC m=+151.569701345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.054678 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.054941 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.554935179 +0000 UTC m=+151.570065305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.122296 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.134338 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.141817 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.163574 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.163802 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.163822 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.163923 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.663909869 +0000 UTC m=+151.679039985 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.268931 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270074 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270152 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psvt5\" (UniqueName: \"kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270200 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270225 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270266 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270289 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.270396 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.270768 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.770750598 +0000 UTC m=+151.785880724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.278306 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.302166 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.314065 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.341670 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.363168 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.371652 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372071 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372093 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372131 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372146 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372204 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76t5j\" (UniqueName: \"kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372244 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psvt5\" (UniqueName: \"kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.372601 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:47.872587312 +0000 UTC m=+151.887717438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372694 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.372874 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.428232 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psvt5\" (UniqueName: \"kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5\") pod \"certified-operators-mfnw5\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.477139 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.477585 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.477654 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.477864 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76t5j\" (UniqueName: \"kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.495559 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.506894 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.507123 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.507407 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.007393848 +0000 UTC m=+152.022523974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.518048 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.518101 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.525843 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.541395 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76t5j\" (UniqueName: \"kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j\") pod \"community-operators-dlzds\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.579839 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.580058 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.580103 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.580155 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm4tk\" (UniqueName: \"kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.580258 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.080242224 +0000 UTC m=+152.095372340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.604181 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.611810 4993 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.621488 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rm9xh" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.681050 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm4tk\" (UniqueName: \"kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.681099 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.681137 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.681154 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.682221 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.682396 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.182380297 +0000 UTC m=+152.197510423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.682583 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.700292 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.748146 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm4tk\" (UniqueName: \"kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk\") pod \"community-operators-7lksb\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.781761 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.782383 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.282333876 +0000 UTC m=+152.297464002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.822526 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b09bdc2ce73be10f7a7a26f8c7fff2f985b6f91e955f8b9578e9bed466d171d8"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.822813 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d292dba28737baf26ec0d172cf9a76f3d7a368981375cc17efcd9451240e8dc0"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.823871 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7ec85d078602f56e657b7bb2870195aeaea52e4b03b15c52f582d447fba9b8a8"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.825056 4993 generic.go:334] "Generic (PLEG): container finished" podID="8835a1b6-a97d-48c8-9dd3-c197250e4825" containerID="ccfc6e9565d9325d95e7125b923c9969c9c76b944792c576c6e12306ce990dd4" exitCode=0 Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.825094 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" event={"ID":"8835a1b6-a97d-48c8-9dd3-c197250e4825","Type":"ContainerDied","Data":"ccfc6e9565d9325d95e7125b923c9969c9c76b944792c576c6e12306ce990dd4"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.849691 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" event={"ID":"b6b4beb5-c112-470d-9041-85f5ebd7cc42","Type":"ContainerStarted","Data":"5b39c89011aaa63e1e55c5ac8178a91f2d171dc332d569e59836b7278d5114db"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.861356 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerStarted","Data":"bc5ff061fe0a5e3fb80564528dd2ed4e2b9da66e251523b7018c007038146ab3"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.873974 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.874927 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5d0d4ab04645bc0103c07f111c2cbc55b2a7e497b7ac44c155ae14bd5bb75642"} Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.883013 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.883935 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.383922163 +0000 UTC m=+152.399052279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:47 crc kubenswrapper[4993]: I0129 12:07:47.984372 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:47 crc kubenswrapper[4993]: E0129 12:07:47.984747 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.484730357 +0000 UTC m=+152.499860473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.014427 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:48 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:48 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:48 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.016341 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.092001 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: E0129 12:07:48.092738 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.592720148 +0000 UTC m=+152.607850274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.151685 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.158551 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.158611 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.193655 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:48 crc kubenswrapper[4993]: E0129 12:07:48.193909 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.693894773 +0000 UTC m=+152.709024889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.281453 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.299469 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:07:48 crc kubenswrapper[4993]: W0129 12:07:48.300950 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7c26ff5_7db3_40d9_bc94_56ee28d89dcb.slice/crio-afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984 WatchSource:0}: Error finding container afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984: Status 404 returned error can't find the container with id afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984 Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.301339 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: E0129 12:07:48.301765 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 12:07:48.801749521 +0000 UTC m=+152.816879647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jz5vk" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.360260 4993 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-29T12:07:47.611840578Z","Handler":null,"Name":""} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.372552 4993 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.372584 4993 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.382620 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.415990 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.438685 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.518679 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.540475 4993 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.540546 4993 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.636354 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.657647 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.657691 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.659613 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-klxb6" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.663220 4993 patch_prober.go:28] interesting pod/console-f9d7485db-pbdv4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.663292 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pbdv4" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.667125 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.691493 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.691552 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.691494 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.691604 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.712672 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.756755 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jz5vk\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.903585 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" event={"ID":"b6b4beb5-c112-470d-9041-85f5ebd7cc42","Type":"ContainerStarted","Data":"63483c91682bbf9f02079ddd3d801cc130069d605c83a40310678b0008d7b3bd"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.911627 4993 generic.go:334] "Generic (PLEG): container finished" podID="98f5acfe-223f-459b-a995-ae89a695ac32" containerID="4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d" exitCode=0 Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.912092 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerDied","Data":"4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.914227 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fc89593-9e3e-4d75-9f87-fed323d7dfbf","Type":"ContainerStarted","Data":"09595cb26ba21875168c6ce513b74f99725c04e77d5a50f1fdaffa46a114058a"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.914389 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.915001 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.931933 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-m8gjn" podStartSLOduration=12.931917557 podStartE2EDuration="12.931917557s" podCreationTimestamp="2026-01-29 12:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:48.930746414 +0000 UTC m=+152.945876560" watchObservedRunningTime="2026-01-29 12:07:48.931917557 +0000 UTC m=+152.947047683" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.934921 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9099c87b581ae56bd60fa03db57359544b0348ba2c246a10832bf5fdf3550c88"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.935131 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.947005 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerStarted","Data":"a1c7b272cf2cee25115eb0b1645af5760721a83f95f298e6465f722c91a3dad2"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.957279 4993 generic.go:334] "Generic (PLEG): container finished" podID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerID="23bc79e9d11a23d95c564334222f47db12214b3607bfa71c67e6b7bdb2bf00d2" exitCode=0 Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.957380 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerDied","Data":"23bc79e9d11a23d95c564334222f47db12214b3607bfa71c67e6b7bdb2bf00d2"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.957405 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerStarted","Data":"afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.970063 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"68763936fec69e6f77aa34e25a874724fa27f64c044b1a5827fe67fc87a6603b"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.975508 4993 generic.go:334] "Generic (PLEG): container finished" podID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerID="a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f" exitCode=0 Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.975723 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerDied","Data":"a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f"} Jan 29 12:07:48 crc kubenswrapper[4993]: I0129 12:07:48.975873 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerStarted","Data":"4173dd18dd23f980a439777fa1ed539edad0666215f618dac54feb54f0a70630"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.009916 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.015977 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:49 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:49 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:49 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.016063 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.050816 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.051756 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.055627 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.097464 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.134926 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.134973 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7czd\" (UniqueName: \"kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.135015 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.183663 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.183695 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.188694 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.198519 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.199563 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.199737 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.204252 4993 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kc9dp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]log ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]etcd ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/generic-apiserver-start-informers ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/max-in-flight-filter ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 29 12:07:49 crc kubenswrapper[4993]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 29 12:07:49 crc kubenswrapper[4993]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/project.openshift.io-projectcache ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/openshift.io-startinformers ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 29 12:07:49 crc kubenswrapper[4993]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 29 12:07:49 crc kubenswrapper[4993]: livez check failed Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.204307 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" podUID="38b7a4c3-bead-47c0-a37a-b0bf359b20aa" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.236376 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.236422 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7czd\" (UniqueName: \"kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.236456 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.238354 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.240643 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.272692 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.290057 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7czd\" (UniqueName: \"kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd\") pod \"redhat-marketplace-n7xrp\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.388032 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.422535 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.444128 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.458401 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.459302 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p67p4"] Jan 29 12:07:49 crc kubenswrapper[4993]: E0129 12:07:49.460969 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8835a1b6-a97d-48c8-9dd3-c197250e4825" containerName="collect-profiles" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.461021 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="8835a1b6-a97d-48c8-9dd3-c197250e4825" containerName="collect-profiles" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.461301 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="8835a1b6-a97d-48c8-9dd3-c197250e4825" containerName="collect-profiles" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.463999 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-27cpg" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.464117 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.478226 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p67p4"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.483473 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7nkd9" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.540792 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume\") pod \"8835a1b6-a97d-48c8-9dd3-c197250e4825\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.540942 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zxbc\" (UniqueName: \"kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc\") pod \"8835a1b6-a97d-48c8-9dd3-c197250e4825\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.541777 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume\") pod \"8835a1b6-a97d-48c8-9dd3-c197250e4825\" (UID: \"8835a1b6-a97d-48c8-9dd3-c197250e4825\") " Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.542073 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjvrz\" (UniqueName: \"kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.542113 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.542136 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.544291 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume" (OuterVolumeSpecName: "config-volume") pod "8835a1b6-a97d-48c8-9dd3-c197250e4825" (UID: "8835a1b6-a97d-48c8-9dd3-c197250e4825"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.549060 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8835a1b6-a97d-48c8-9dd3-c197250e4825" (UID: "8835a1b6-a97d-48c8-9dd3-c197250e4825"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.551602 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc" (OuterVolumeSpecName: "kube-api-access-4zxbc") pod "8835a1b6-a97d-48c8-9dd3-c197250e4825" (UID: "8835a1b6-a97d-48c8-9dd3-c197250e4825"). InnerVolumeSpecName "kube-api-access-4zxbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644071 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjvrz\" (UniqueName: \"kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644125 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644155 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644230 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zxbc\" (UniqueName: \"kubernetes.io/projected/8835a1b6-a97d-48c8-9dd3-c197250e4825-kube-api-access-4zxbc\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644248 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8835a1b6-a97d-48c8-9dd3-c197250e4825-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644262 4993 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8835a1b6-a97d-48c8-9dd3-c197250e4825-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.644807 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.645162 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.655693 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.657121 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.660796 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.661392 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.665830 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.677525 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjvrz\" (UniqueName: \"kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz\") pod \"redhat-marketplace-p67p4\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.709763 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.712945 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.721657 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.728425 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qspck" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.744917 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.747012 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.812828 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.848394 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.848536 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.849028 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.869588 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.986441 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.990208 4993 generic.go:334] "Generic (PLEG): container finished" podID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerID="091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5" exitCode=0 Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.990275 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerDied","Data":"091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.990303 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerStarted","Data":"ec25f9dbb35be43ce5425a9dd7a136c92ee52a22545d828ff89d77eba70ea1af"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.993702 4993 generic.go:334] "Generic (PLEG): container finished" podID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerID="95d35d33904b9b205a0fc1fb618988c97df99fd87d1f405a82e6d840f8bca933" exitCode=0 Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.993744 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerDied","Data":"95d35d33904b9b205a0fc1fb618988c97df99fd87d1f405a82e6d840f8bca933"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.996756 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" event={"ID":"0e93bec6-c6d1-4018-8da3-73f918e4d5b6","Type":"ContainerStarted","Data":"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.996832 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" event={"ID":"0e93bec6-c6d1-4018-8da3-73f918e4d5b6","Type":"ContainerStarted","Data":"6376f74ddec262b10a1930ceb69be03cde928f084c7de903a4b2c4a52474bb3f"} Jan 29 12:07:49 crc kubenswrapper[4993]: I0129 12:07:49.996854 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:49.999960 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.000425 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb" event={"ID":"8835a1b6-a97d-48c8-9dd3-c197250e4825","Type":"ContainerDied","Data":"99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5"} Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.000456 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99a154a287c37f33a0d0ae6744ef9f5313683728316e2db698381f5f48cb69d5" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.003166 4993 generic.go:334] "Generic (PLEG): container finished" podID="9fc89593-9e3e-4d75-9f87-fed323d7dfbf" containerID="961e78906f62598dcdb0f2d47b9cb04ca70cc27d1c2feea03883ab446189c300" exitCode=0 Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.003384 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fc89593-9e3e-4d75-9f87-fed323d7dfbf","Type":"ContainerDied","Data":"961e78906f62598dcdb0f2d47b9cb04ca70cc27d1c2feea03883ab446189c300"} Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.017681 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:50 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:50 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:50 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.017736 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.018424 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjhzh" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.056756 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.057934 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.070001 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.076243 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" podStartSLOduration=133.076216977 podStartE2EDuration="2m13.076216977s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:07:50.04310669 +0000 UTC m=+154.058236836" watchObservedRunningTime="2026-01-29 12:07:50.076216977 +0000 UTC m=+154.091347203" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.102393 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.153955 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.154031 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjpsj\" (UniqueName: \"kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.154212 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.255808 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.256204 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.256249 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjpsj\" (UniqueName: \"kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.256483 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.256651 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.287542 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjpsj\" (UniqueName: \"kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj\") pod \"redhat-operators-9nh5h\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.342381 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p67p4"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.382899 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.420498 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.453422 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.454639 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.455659 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.564009 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.564065 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkgjf\" (UniqueName: \"kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.564124 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.664738 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkgjf\" (UniqueName: \"kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.665098 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.665154 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.665650 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.666174 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.701785 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkgjf\" (UniqueName: \"kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf\") pod \"redhat-operators-86rvn\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.797227 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:07:50 crc kubenswrapper[4993]: I0129 12:07:50.892715 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:07:51 crc kubenswrapper[4993]: W0129 12:07:51.622057 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60d2ac3a_4a78_4bad_af91_633f78900181.slice/crio-54ac48cebd9efded5c8d3bc2a9eca90b236dbd65f862ec09a07906a40af63a97 WatchSource:0}: Error finding container 54ac48cebd9efded5c8d3bc2a9eca90b236dbd65f862ec09a07906a40af63a97: Status 404 returned error can't find the container with id 54ac48cebd9efded5c8d3bc2a9eca90b236dbd65f862ec09a07906a40af63a97 Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.629089 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:51 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:51 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:51 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.629137 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.693842 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7fef86d0-ff61-46b0-b34e-071d590b17fe","Type":"ContainerStarted","Data":"c8092d456eb629c89bb32020fcbcaf3793de0f823b9730b20638e37d5004bff7"} Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.702331 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerStarted","Data":"54ac48cebd9efded5c8d3bc2a9eca90b236dbd65f862ec09a07906a40af63a97"} Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.723815 4993 generic.go:334] "Generic (PLEG): container finished" podID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerID="b16b2fdd35435f6dfb078fb999f07ffd5844d3b82f291cfc0fc82592d4c3f6f4" exitCode=0 Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.724414 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerDied","Data":"b16b2fdd35435f6dfb078fb999f07ffd5844d3b82f291cfc0fc82592d4c3f6f4"} Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.724463 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerStarted","Data":"874b0c2523b59170414f12b384e255f4bf2d60461f3450b44c5f99517a87403f"} Jan 29 12:07:51 crc kubenswrapper[4993]: I0129 12:07:51.890542 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.014324 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:52 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:52 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:52 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.014767 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.122343 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.249776 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir\") pod \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.249948 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access\") pod \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\" (UID: \"9fc89593-9e3e-4d75-9f87-fed323d7dfbf\") " Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.250230 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9fc89593-9e3e-4d75-9f87-fed323d7dfbf" (UID: "9fc89593-9e3e-4d75-9f87-fed323d7dfbf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.250502 4993 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.259814 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9fc89593-9e3e-4d75-9f87-fed323d7dfbf" (UID: "9fc89593-9e3e-4d75-9f87-fed323d7dfbf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.352071 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fc89593-9e3e-4d75-9f87-fed323d7dfbf-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.734361 4993 generic.go:334] "Generic (PLEG): container finished" podID="7fef86d0-ff61-46b0-b34e-071d590b17fe" containerID="075d721379575753ac827fe551e6f0af6249c7fab4925abaeea6b9eb4982cfe4" exitCode=0 Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.734447 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7fef86d0-ff61-46b0-b34e-071d590b17fe","Type":"ContainerDied","Data":"075d721379575753ac827fe551e6f0af6249c7fab4925abaeea6b9eb4982cfe4"} Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.738378 4993 generic.go:334] "Generic (PLEG): container finished" podID="af46a836-d45e-493b-9503-40281a2149d5" containerID="855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c" exitCode=0 Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.738472 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerDied","Data":"855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c"} Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.738525 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerStarted","Data":"a1e516b6e0706fb6b42a33c12b7ac926b650d9694f29579ca5a706c7bed059a6"} Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.741336 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fc89593-9e3e-4d75-9f87-fed323d7dfbf","Type":"ContainerDied","Data":"09595cb26ba21875168c6ce513b74f99725c04e77d5a50f1fdaffa46a114058a"} Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.741371 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09595cb26ba21875168c6ce513b74f99725c04e77d5a50f1fdaffa46a114058a" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.741436 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.761400 4993 generic.go:334] "Generic (PLEG): container finished" podID="60d2ac3a-4a78-4bad-af91-633f78900181" containerID="d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e" exitCode=0 Jan 29 12:07:52 crc kubenswrapper[4993]: I0129 12:07:52.761720 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerDied","Data":"d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e"} Jan 29 12:07:53 crc kubenswrapper[4993]: I0129 12:07:53.011920 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:53 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:53 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:53 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:53 crc kubenswrapper[4993]: I0129 12:07:53.011978 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.030545 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:54 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:54 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:54 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.030860 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.035209 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.081221 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access\") pod \"7fef86d0-ff61-46b0-b34e-071d590b17fe\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.082212 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir\") pod \"7fef86d0-ff61-46b0-b34e-071d590b17fe\" (UID: \"7fef86d0-ff61-46b0-b34e-071d590b17fe\") " Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.082348 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7fef86d0-ff61-46b0-b34e-071d590b17fe" (UID: "7fef86d0-ff61-46b0-b34e-071d590b17fe"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.082674 4993 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7fef86d0-ff61-46b0-b34e-071d590b17fe-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.100365 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7fef86d0-ff61-46b0-b34e-071d590b17fe" (UID: "7fef86d0-ff61-46b0-b34e-071d590b17fe"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.185932 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fef86d0-ff61-46b0-b34e-071d590b17fe-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.190402 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.198980 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kc9dp" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.698070 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-79rwf" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.787268 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.787255 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7fef86d0-ff61-46b0-b34e-071d590b17fe","Type":"ContainerDied","Data":"c8092d456eb629c89bb32020fcbcaf3793de0f823b9730b20638e37d5004bff7"} Jan 29 12:07:54 crc kubenswrapper[4993]: I0129 12:07:54.787303 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8092d456eb629c89bb32020fcbcaf3793de0f823b9730b20638e37d5004bff7" Jan 29 12:07:55 crc kubenswrapper[4993]: I0129 12:07:55.013566 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:55 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:55 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:55 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:55 crc kubenswrapper[4993]: I0129 12:07:55.013615 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:56 crc kubenswrapper[4993]: I0129 12:07:56.016042 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:56 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:56 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:56 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:56 crc kubenswrapper[4993]: I0129 12:07:56.016104 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:57 crc kubenswrapper[4993]: I0129 12:07:57.013968 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:57 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:57 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:57 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:57 crc kubenswrapper[4993]: I0129 12:07:57.014275 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.013670 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:58 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:58 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:58 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.013750 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.657636 4993 patch_prober.go:28] interesting pod/console-f9d7485db-pbdv4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.657702 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pbdv4" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.691455 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.691483 4993 patch_prober.go:28] interesting pod/downloads-7954f5f757-mwxh6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.691515 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:58 crc kubenswrapper[4993]: I0129 12:07:58.691515 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mwxh6" podUID="b1ec8778-2a97-490f-a822-15d8bd226643" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 12:07:59 crc kubenswrapper[4993]: I0129 12:07:59.012536 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:07:59 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:07:59 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:07:59 crc kubenswrapper[4993]: healthz check failed Jan 29 12:07:59 crc kubenswrapper[4993]: I0129 12:07:59.012591 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:00 crc kubenswrapper[4993]: I0129 12:08:00.013174 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:00 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:08:00 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:00 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:00 crc kubenswrapper[4993]: I0129 12:08:00.013527 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:01 crc kubenswrapper[4993]: I0129 12:08:01.011976 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:01 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:08:01 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:01 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:01 crc kubenswrapper[4993]: I0129 12:08:01.012039 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:01 crc kubenswrapper[4993]: I0129 12:08:01.293509 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:08:01 crc kubenswrapper[4993]: I0129 12:08:01.303910 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7838e5a2-ff4a-42b8-aeef-fefbe1176fc4-metrics-certs\") pod \"network-metrics-daemon-vdgbx\" (UID: \"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4\") " pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:08:01 crc kubenswrapper[4993]: I0129 12:08:01.315632 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vdgbx" Jan 29 12:08:02 crc kubenswrapper[4993]: I0129 12:08:02.013545 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:02 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:08:02 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:02 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:02 crc kubenswrapper[4993]: I0129 12:08:02.013830 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:03 crc kubenswrapper[4993]: I0129 12:08:03.012988 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:03 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:08:03 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:03 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:03 crc kubenswrapper[4993]: I0129 12:08:03.014024 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:04 crc kubenswrapper[4993]: I0129 12:08:04.012141 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:04 crc kubenswrapper[4993]: [-]has-synced failed: reason withheld Jan 29 12:08:04 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:04 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:04 crc kubenswrapper[4993]: I0129 12:08:04.012497 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.013660 4993 patch_prober.go:28] interesting pod/router-default-5444994796-s4sgr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 12:08:05 crc kubenswrapper[4993]: [+]has-synced ok Jan 29 12:08:05 crc kubenswrapper[4993]: [+]process-running ok Jan 29 12:08:05 crc kubenswrapper[4993]: healthz check failed Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.013720 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s4sgr" podUID="0517259c-9701-45de-9b5e-a71fb348fc2b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.816010 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.816925 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" containerID="cri-o://add21d2a94271a16ec18c6721c654ee230f731dbf165d327132696d337ec7961" gracePeriod=30 Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.823680 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:08:05 crc kubenswrapper[4993]: I0129 12:08:05.824145 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" containerID="cri-o://7bc9e25c72d2ec3b3197ca3b756a667d21b97217ad8ba2d1de0c0aeaf2bb176d" gracePeriod=30 Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.011995 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.014617 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-s4sgr" Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.891484 4993 generic.go:334] "Generic (PLEG): container finished" podID="d62f8973-63e6-4941-aa65-442c43e23b02" containerID="add21d2a94271a16ec18c6721c654ee230f731dbf165d327132696d337ec7961" exitCode=0 Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.891541 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" event={"ID":"d62f8973-63e6-4941-aa65-442c43e23b02","Type":"ContainerDied","Data":"add21d2a94271a16ec18c6721c654ee230f731dbf165d327132696d337ec7961"} Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.893472 4993 generic.go:334] "Generic (PLEG): container finished" podID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerID="7bc9e25c72d2ec3b3197ca3b756a667d21b97217ad8ba2d1de0c0aeaf2bb176d" exitCode=0 Jan 29 12:08:06 crc kubenswrapper[4993]: I0129 12:08:06.893568 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" event={"ID":"b7559669-0e20-491a-95ec-3beacab1c1e1","Type":"ContainerDied","Data":"7bc9e25c72d2ec3b3197ca3b756a667d21b97217ad8ba2d1de0c0aeaf2bb176d"} Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.267996 4993 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kqng8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.268057 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.657548 4993 patch_prober.go:28] interesting pod/console-f9d7485db-pbdv4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.657604 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pbdv4" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.684735 4993 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-md8vs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.684781 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.697464 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mwxh6" Jan 29 12:08:08 crc kubenswrapper[4993]: I0129 12:08:08.921220 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:08:17 crc kubenswrapper[4993]: E0129 12:08:17.366554 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 12:08:17 crc kubenswrapper[4993]: E0129 12:08:17.367169 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-psvt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mfnw5_openshift-marketplace(f7c26ff5-7db3-40d9-bc94-56ee28d89dcb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:17 crc kubenswrapper[4993]: E0129 12:08:17.368299 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mfnw5" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" Jan 29 12:08:18 crc kubenswrapper[4993]: I0129 12:08:18.156580 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:08:18 crc kubenswrapper[4993]: I0129 12:08:18.156645 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:08:18 crc kubenswrapper[4993]: I0129 12:08:18.664520 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:08:18 crc kubenswrapper[4993]: I0129 12:08:18.671842 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:08:19 crc kubenswrapper[4993]: I0129 12:08:19.267274 4993 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kqng8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 12:08:19 crc kubenswrapper[4993]: I0129 12:08:19.267789 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 12:08:19 crc kubenswrapper[4993]: I0129 12:08:19.684999 4993 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-md8vs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 12:08:19 crc kubenswrapper[4993]: I0129 12:08:19.685125 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 12:08:19 crc kubenswrapper[4993]: I0129 12:08:19.715416 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j4tp6" Jan 29 12:08:24 crc kubenswrapper[4993]: E0129 12:08:24.051492 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mfnw5" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.840029 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 12:08:26 crc kubenswrapper[4993]: E0129 12:08:26.840576 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fef86d0-ff61-46b0-b34e-071d590b17fe" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.840588 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fef86d0-ff61-46b0-b34e-071d590b17fe" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: E0129 12:08:26.840601 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc89593-9e3e-4d75-9f87-fed323d7dfbf" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.840606 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc89593-9e3e-4d75-9f87-fed323d7dfbf" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.840698 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fef86d0-ff61-46b0-b34e-071d590b17fe" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.840711 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc89593-9e3e-4d75-9f87-fed323d7dfbf" containerName="pruner" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.841061 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.843377 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.843608 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.853058 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.956070 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:26 crc kubenswrapper[4993]: I0129 12:08:26.956145 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.057815 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.057886 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.057929 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.069757 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.094771 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:27 crc kubenswrapper[4993]: I0129 12:08:27.158569 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:29 crc kubenswrapper[4993]: I0129 12:08:29.268234 4993 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kqng8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 12:08:29 crc kubenswrapper[4993]: I0129 12:08:29.268616 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 12:08:29 crc kubenswrapper[4993]: I0129 12:08:29.684949 4993 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-md8vs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 12:08:29 crc kubenswrapper[4993]: I0129 12:08:29.685038 4993 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.840726 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.848934 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.849271 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.903502 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.903597 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:30 crc kubenswrapper[4993]: I0129 12:08:30.903768 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:30 crc kubenswrapper[4993]: E0129 12:08:30.976341 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 12:08:30 crc kubenswrapper[4993]: E0129 12:08:30.976758 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tjpsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9nh5h_openshift-marketplace(60d2ac3a-4a78-4bad-af91-633f78900181): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:30 crc kubenswrapper[4993]: E0129 12:08:30.978010 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9nh5h" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.004540 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.004609 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.004634 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.004696 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.004738 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.024299 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access\") pod \"installer-9-crc\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:31 crc kubenswrapper[4993]: E0129 12:08:31.113697 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 12:08:31 crc kubenswrapper[4993]: E0129 12:08:31.113854 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-76t5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-dlzds_openshift-marketplace(afa6fe4a-b638-4af7-9b26-5d2208d4e185): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:31 crc kubenswrapper[4993]: E0129 12:08:31.115203 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-dlzds" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" Jan 29 12:08:31 crc kubenswrapper[4993]: I0129 12:08:31.172319 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:08:33 crc kubenswrapper[4993]: E0129 12:08:33.662389 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9nh5h" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" Jan 29 12:08:33 crc kubenswrapper[4993]: E0129 12:08:33.662879 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-dlzds" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.770391 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.783016 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.816477 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:08:33 crc kubenswrapper[4993]: E0129 12:08:33.816699 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.816711 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: E0129 12:08:33.816722 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.816728 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.816837 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" containerName="controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.816853 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" containerName="route-controller-manager" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.817290 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.828677 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.840953 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert\") pod \"b7559669-0e20-491a-95ec-3beacab1c1e1\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.841060 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szjgz\" (UniqueName: \"kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz\") pod \"b7559669-0e20-491a-95ec-3beacab1c1e1\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.841143 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles\") pod \"d62f8973-63e6-4941-aa65-442c43e23b02\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842403 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-996g7\" (UniqueName: \"kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7\") pod \"d62f8973-63e6-4941-aa65-442c43e23b02\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842434 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config\") pod \"b7559669-0e20-491a-95ec-3beacab1c1e1\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842465 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca\") pod \"b7559669-0e20-491a-95ec-3beacab1c1e1\" (UID: \"b7559669-0e20-491a-95ec-3beacab1c1e1\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842507 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca\") pod \"d62f8973-63e6-4941-aa65-442c43e23b02\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842541 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config\") pod \"d62f8973-63e6-4941-aa65-442c43e23b02\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842570 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert\") pod \"d62f8973-63e6-4941-aa65-442c43e23b02\" (UID: \"d62f8973-63e6-4941-aa65-442c43e23b02\") " Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842800 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842826 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842875 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.842961 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv7lv\" (UniqueName: \"kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.846330 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config" (OuterVolumeSpecName: "config") pod "d62f8973-63e6-4941-aa65-442c43e23b02" (UID: "d62f8973-63e6-4941-aa65-442c43e23b02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.847246 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config" (OuterVolumeSpecName: "config") pod "b7559669-0e20-491a-95ec-3beacab1c1e1" (UID: "b7559669-0e20-491a-95ec-3beacab1c1e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.847404 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca" (OuterVolumeSpecName: "client-ca") pod "b7559669-0e20-491a-95ec-3beacab1c1e1" (UID: "b7559669-0e20-491a-95ec-3beacab1c1e1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.847764 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca" (OuterVolumeSpecName: "client-ca") pod "d62f8973-63e6-4941-aa65-442c43e23b02" (UID: "d62f8973-63e6-4941-aa65-442c43e23b02"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.850585 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d62f8973-63e6-4941-aa65-442c43e23b02" (UID: "d62f8973-63e6-4941-aa65-442c43e23b02"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.851480 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7" (OuterVolumeSpecName: "kube-api-access-996g7") pod "d62f8973-63e6-4941-aa65-442c43e23b02" (UID: "d62f8973-63e6-4941-aa65-442c43e23b02"). InnerVolumeSpecName "kube-api-access-996g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.849760 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d62f8973-63e6-4941-aa65-442c43e23b02" (UID: "d62f8973-63e6-4941-aa65-442c43e23b02"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.862367 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b7559669-0e20-491a-95ec-3beacab1c1e1" (UID: "b7559669-0e20-491a-95ec-3beacab1c1e1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.863917 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz" (OuterVolumeSpecName: "kube-api-access-szjgz") pod "b7559669-0e20-491a-95ec-3beacab1c1e1" (UID: "b7559669-0e20-491a-95ec-3beacab1c1e1"). InnerVolumeSpecName "kube-api-access-szjgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.901645 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vdgbx"] Jan 29 12:08:33 crc kubenswrapper[4993]: W0129 12:08:33.910706 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7838e5a2_ff4a_42b8_aeef_fefbe1176fc4.slice/crio-abe846dd26713afd0cb737823bc0a2c1c5c077313e414863e33b95ac60631cc9 WatchSource:0}: Error finding container abe846dd26713afd0cb737823bc0a2c1c5c077313e414863e33b95ac60631cc9: Status 404 returned error can't find the container with id abe846dd26713afd0cb737823bc0a2c1c5c077313e414863e33b95ac60631cc9 Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961802 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961843 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961873 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961913 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv7lv\" (UniqueName: \"kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961959 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961968 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d62f8973-63e6-4941-aa65-442c43e23b02-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961978 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7559669-0e20-491a-95ec-3beacab1c1e1-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961987 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szjgz\" (UniqueName: \"kubernetes.io/projected/b7559669-0e20-491a-95ec-3beacab1c1e1-kube-api-access-szjgz\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.961996 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.962003 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-996g7\" (UniqueName: \"kubernetes.io/projected/d62f8973-63e6-4941-aa65-442c43e23b02-kube-api-access-996g7\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.962013 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.962020 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7559669-0e20-491a-95ec-3beacab1c1e1-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.962030 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d62f8973-63e6-4941-aa65-442c43e23b02-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.965265 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.966998 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:33 crc kubenswrapper[4993]: I0129 12:08:33.997638 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv7lv\" (UniqueName: \"kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.019424 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 12:08:34 crc kubenswrapper[4993]: W0129 12:08:34.028527 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod825bd7fd_af79_43b0_9aef_28a7ee9a2417.slice/crio-3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a WatchSource:0}: Error finding container 3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a: Status 404 returned error can't find the container with id 3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.028821 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.028858 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-md8vs" event={"ID":"d62f8973-63e6-4941-aa65-442c43e23b02","Type":"ContainerDied","Data":"cc5ce3d2b61d346f212fe3fdde511a1df9d9e683f759f3946c6944b238abb8b0"} Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.028939 4993 scope.go:117] "RemoveContainer" containerID="add21d2a94271a16ec18c6721c654ee230f731dbf165d327132696d337ec7961" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.031597 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" event={"ID":"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4","Type":"ContainerStarted","Data":"abe846dd26713afd0cb737823bc0a2c1c5c077313e414863e33b95ac60631cc9"} Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.035115 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" event={"ID":"b7559669-0e20-491a-95ec-3beacab1c1e1","Type":"ContainerDied","Data":"4685306264dbc9f62ae1e7d3c33448c3fecbc4bde2fa144b4ec2db8e80b641ab"} Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.035262 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.070580 4993 scope.go:117] "RemoveContainer" containerID="7bc9e25c72d2ec3b3197ca3b756a667d21b97217ad8ba2d1de0c0aeaf2bb176d" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.074695 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.082917 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.087123 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kqng8"] Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.091173 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.093663 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-md8vs"] Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.171437 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config\") pod \"route-controller-manager-b77dc47b6-vr2td\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:34 crc kubenswrapper[4993]: W0129 12:08:34.181141 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod39088c04_e684_40fc_bdec_8933958c2a4e.slice/crio-9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a WatchSource:0}: Error finding container 9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a: Status 404 returned error can't find the container with id 9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.188037 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:34 crc kubenswrapper[4993]: I0129 12:08:34.442460 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:08:34 crc kubenswrapper[4993]: W0129 12:08:34.514965 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ec74443_272b_4f3e_b4f8_8cf5c62db9a6.slice/crio-bab02233140d9afa82f2964e716ee0a83e49665e155a9e5c3566ff9156aba45e WatchSource:0}: Error finding container bab02233140d9afa82f2964e716ee0a83e49665e155a9e5c3566ff9156aba45e: Status 404 returned error can't find the container with id bab02233140d9afa82f2964e716ee0a83e49665e155a9e5c3566ff9156aba45e Jan 29 12:08:35 crc kubenswrapper[4993]: I0129 12:08:35.043867 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"825bd7fd-af79-43b0-9aef-28a7ee9a2417","Type":"ContainerStarted","Data":"3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a"} Jan 29 12:08:35 crc kubenswrapper[4993]: I0129 12:08:35.044882 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"39088c04-e684-40fc-bdec-8933958c2a4e","Type":"ContainerStarted","Data":"9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a"} Jan 29 12:08:35 crc kubenswrapper[4993]: I0129 12:08:35.046661 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" event={"ID":"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6","Type":"ContainerStarted","Data":"bab02233140d9afa82f2964e716ee0a83e49665e155a9e5c3566ff9156aba45e"} Jan 29 12:08:35 crc kubenswrapper[4993]: I0129 12:08:35.197489 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7559669-0e20-491a-95ec-3beacab1c1e1" path="/var/lib/kubelet/pods/b7559669-0e20-491a-95ec-3beacab1c1e1/volumes" Jan 29 12:08:35 crc kubenswrapper[4993]: I0129 12:08:35.198709 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62f8973-63e6-4941-aa65-442c43e23b02" path="/var/lib/kubelet/pods/d62f8973-63e6-4941-aa65-442c43e23b02/volumes" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.665081 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.666126 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.668413 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.668495 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.668413 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.668650 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.669170 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.669226 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.676516 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.683760 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.704975 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vznjm\" (UniqueName: \"kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.705114 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.705172 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.705343 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.705390 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.807291 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vznjm\" (UniqueName: \"kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.807424 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.807459 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.808421 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.808474 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.809066 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.809515 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.809872 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.818706 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:36 crc kubenswrapper[4993]: I0129 12:08:36.824056 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vznjm\" (UniqueName: \"kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm\") pod \"controller-manager-6bfb7f7847-cjs2d\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:37 crc kubenswrapper[4993]: I0129 12:08:37.047904 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:37 crc kubenswrapper[4993]: I0129 12:08:37.240356 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:08:37 crc kubenswrapper[4993]: W0129 12:08:37.300732 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bfb01f8_c2bf_4975_9fdc_bbf303c00295.slice/crio-2e298a73fb6a7916d73cbff7f622449b3b723448609776381205af5b307a4838 WatchSource:0}: Error finding container 2e298a73fb6a7916d73cbff7f622449b3b723448609776381205af5b307a4838: Status 404 returned error can't find the container with id 2e298a73fb6a7916d73cbff7f622449b3b723448609776381205af5b307a4838 Jan 29 12:08:37 crc kubenswrapper[4993]: E0129 12:08:37.764287 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 12:08:37 crc kubenswrapper[4993]: E0129 12:08:37.764432 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hjvrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p67p4_openshift-marketplace(8421bbe0-da5e-49ec-994f-fd9a072a4bc7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:37 crc kubenswrapper[4993]: E0129 12:08:37.765658 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p67p4" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.067269 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"825bd7fd-af79-43b0-9aef-28a7ee9a2417","Type":"ContainerStarted","Data":"b88811e913c314542ac3d0ccd7163433b1c28c8f5eb0d50165f37260114fb521"} Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.071078 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"39088c04-e684-40fc-bdec-8933958c2a4e","Type":"ContainerStarted","Data":"c3af66244ef21d53950a1f1d1867e20c93215d02e78feee392ac0d69f38f0c73"} Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.078664 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" event={"ID":"4bfb01f8-c2bf-4975-9fdc-bbf303c00295","Type":"ContainerStarted","Data":"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01"} Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.078743 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" event={"ID":"4bfb01f8-c2bf-4975-9fdc-bbf303c00295","Type":"ContainerStarted","Data":"2e298a73fb6a7916d73cbff7f622449b3b723448609776381205af5b307a4838"} Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.078769 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.080629 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" event={"ID":"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6","Type":"ContainerStarted","Data":"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f"} Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.080867 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.082143 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" event={"ID":"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4","Type":"ContainerStarted","Data":"967bdb3d4f8bee851f70002ad1528e553a54accdc9f5a64df44a7c5fea964ec9"} Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.083403 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p67p4" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.088958 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.099909 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=12.099880597 podStartE2EDuration="12.099880597s" podCreationTimestamp="2026-01-29 12:08:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:08:38.095533177 +0000 UTC m=+202.110663303" watchObservedRunningTime="2026-01-29 12:08:38.099880597 +0000 UTC m=+202.115010723" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.152115 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.152096713 podStartE2EDuration="8.152096713s" podCreationTimestamp="2026-01-29 12:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:08:38.116017134 +0000 UTC m=+202.131147260" watchObservedRunningTime="2026-01-29 12:08:38.152096713 +0000 UTC m=+202.167226839" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.190104 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" podStartSLOduration=13.190089423 podStartE2EDuration="13.190089423s" podCreationTimestamp="2026-01-29 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:08:38.189204794 +0000 UTC m=+202.204334920" watchObservedRunningTime="2026-01-29 12:08:38.190089423 +0000 UTC m=+202.205219549" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.214473 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" podStartSLOduration=13.214454336 podStartE2EDuration="13.214454336s" podCreationTimestamp="2026-01-29 12:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:08:38.212858263 +0000 UTC m=+202.227988389" watchObservedRunningTime="2026-01-29 12:08:38.214454336 +0000 UTC m=+202.229584452" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.366005 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.366175 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qkgjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-86rvn_openshift-marketplace(af46a836-d45e-493b-9503-40281a2149d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.367390 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-86rvn" podUID="af46a836-d45e-493b-9503-40281a2149d5" Jan 29 12:08:38 crc kubenswrapper[4993]: I0129 12:08:38.554208 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.735496 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.735651 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rm4tk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7lksb_openshift-marketplace(ceb99f3c-f71e-4e45-9336-bac8338c94b7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.736907 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7lksb" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.772329 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.772520 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7czd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-n7xrp_openshift-marketplace(9036c1c8-0e08-4924-abee-1d40400d6e96): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:38 crc kubenswrapper[4993]: E0129 12:08:38.773725 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-n7xrp" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" Jan 29 12:08:39 crc kubenswrapper[4993]: I0129 12:08:39.089812 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vdgbx" event={"ID":"7838e5a2-ff4a-42b8-aeef-fefbe1176fc4","Type":"ContainerStarted","Data":"ea7166226ec3d6190aa933e3a5f5d36cd6d246fcaefbb7b8b0c97402307555a6"} Jan 29 12:08:39 crc kubenswrapper[4993]: E0129 12:08:39.091446 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-86rvn" podUID="af46a836-d45e-493b-9503-40281a2149d5" Jan 29 12:08:39 crc kubenswrapper[4993]: E0129 12:08:39.091942 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-n7xrp" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" Jan 29 12:08:39 crc kubenswrapper[4993]: E0129 12:08:39.092085 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7lksb" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" Jan 29 12:08:39 crc kubenswrapper[4993]: I0129 12:08:39.137097 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vdgbx" podStartSLOduration=182.137079778 podStartE2EDuration="3m2.137079778s" podCreationTimestamp="2026-01-29 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:08:39.134436784 +0000 UTC m=+203.149566900" watchObservedRunningTime="2026-01-29 12:08:39.137079778 +0000 UTC m=+203.152209894" Jan 29 12:08:41 crc kubenswrapper[4993]: E0129 12:08:41.754732 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 12:08:41 crc kubenswrapper[4993]: E0129 12:08:41.755218 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qt8jz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zzw5p_openshift-marketplace(98f5acfe-223f-459b-a995-ae89a695ac32): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 12:08:41 crc kubenswrapper[4993]: E0129 12:08:41.757516 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zzw5p" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" Jan 29 12:08:42 crc kubenswrapper[4993]: E0129 12:08:42.104491 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zzw5p" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" Jan 29 12:08:47 crc kubenswrapper[4993]: I0129 12:08:47.136460 4993 generic.go:334] "Generic (PLEG): container finished" podID="825bd7fd-af79-43b0-9aef-28a7ee9a2417" containerID="b88811e913c314542ac3d0ccd7163433b1c28c8f5eb0d50165f37260114fb521" exitCode=0 Jan 29 12:08:47 crc kubenswrapper[4993]: I0129 12:08:47.136550 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"825bd7fd-af79-43b0-9aef-28a7ee9a2417","Type":"ContainerDied","Data":"b88811e913c314542ac3d0ccd7163433b1c28c8f5eb0d50165f37260114fb521"} Jan 29 12:08:48 crc kubenswrapper[4993]: I0129 12:08:48.156938 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:08:48 crc kubenswrapper[4993]: I0129 12:08:48.157027 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:08:48 crc kubenswrapper[4993]: I0129 12:08:48.157109 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:08:48 crc kubenswrapper[4993]: I0129 12:08:48.158153 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:08:48 crc kubenswrapper[4993]: I0129 12:08:48.158797 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018" gracePeriod=600 Jan 29 12:08:51 crc kubenswrapper[4993]: I0129 12:08:51.184302 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018" exitCode=0 Jan 29 12:08:51 crc kubenswrapper[4993]: I0129 12:08:51.184389 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018"} Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.253049 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.324350 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access\") pod \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.324523 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir\") pod \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\" (UID: \"825bd7fd-af79-43b0-9aef-28a7ee9a2417\") " Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.325148 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "825bd7fd-af79-43b0-9aef-28a7ee9a2417" (UID: "825bd7fd-af79-43b0-9aef-28a7ee9a2417"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.331015 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "825bd7fd-af79-43b0-9aef-28a7ee9a2417" (UID: "825bd7fd-af79-43b0-9aef-28a7ee9a2417"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.382904 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"825bd7fd-af79-43b0-9aef-28a7ee9a2417","Type":"ContainerDied","Data":"3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a"} Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.382949 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.382958 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f7f0c809a0f61d0c301e45d7c9f2a536fd36778b6251d35c76ff794a71c7d3a" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.426748 4993 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:53 crc kubenswrapper[4993]: I0129 12:08:53.426890 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/825bd7fd-af79-43b0-9aef-28a7ee9a2417-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:08:54 crc kubenswrapper[4993]: I0129 12:08:54.394397 4993 generic.go:334] "Generic (PLEG): container finished" podID="60d2ac3a-4a78-4bad-af91-633f78900181" containerID="90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2" exitCode=0 Jan 29 12:08:54 crc kubenswrapper[4993]: I0129 12:08:54.394570 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerDied","Data":"90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2"} Jan 29 12:08:54 crc kubenswrapper[4993]: I0129 12:08:54.399111 4993 generic.go:334] "Generic (PLEG): container finished" podID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerID="7ad2efeeca225dc2763a2b1b03b14ec0cd497d735f2c5b3f59a3854732aa3d84" exitCode=0 Jan 29 12:08:54 crc kubenswrapper[4993]: I0129 12:08:54.399151 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerDied","Data":"7ad2efeeca225dc2763a2b1b03b14ec0cd497d735f2c5b3f59a3854732aa3d84"} Jan 29 12:08:54 crc kubenswrapper[4993]: I0129 12:08:54.402432 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.415147 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerStarted","Data":"3904233c3a323ecbf0b9d65c3d54252e4299a85773b27eca90f1b6d12eaf4298"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.416640 4993 generic.go:334] "Generic (PLEG): container finished" podID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerID="6bb1c547ebc5847bf94dba64b05b4b6c5eca4d23254cc736c1890253ade50238" exitCode=0 Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.416695 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerDied","Data":"6bb1c547ebc5847bf94dba64b05b4b6c5eca4d23254cc736c1890253ade50238"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.418825 4993 generic.go:334] "Generic (PLEG): container finished" podID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerID="573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6" exitCode=0 Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.418888 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerDied","Data":"573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.420834 4993 generic.go:334] "Generic (PLEG): container finished" podID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerID="c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0" exitCode=0 Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.420878 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerDied","Data":"c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.423619 4993 generic.go:334] "Generic (PLEG): container finished" podID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerID="eaba1b246b8f7dd6c086020e84b066146f338f16e8c6b7b27d75632fd43a6538" exitCode=0 Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.423681 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerDied","Data":"eaba1b246b8f7dd6c086020e84b066146f338f16e8c6b7b27d75632fd43a6538"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.426493 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerStarted","Data":"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede"} Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.452540 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mfnw5" podStartSLOduration=2.442739151 podStartE2EDuration="1m9.452521343s" podCreationTimestamp="2026-01-29 12:07:47 +0000 UTC" firstStartedPulling="2026-01-29 12:07:48.959276799 +0000 UTC m=+152.974406925" lastFinishedPulling="2026-01-29 12:08:55.969058991 +0000 UTC m=+219.984189117" observedRunningTime="2026-01-29 12:08:56.436986675 +0000 UTC m=+220.452116801" watchObservedRunningTime="2026-01-29 12:08:56.452521343 +0000 UTC m=+220.467651469" Jan 29 12:08:56 crc kubenswrapper[4993]: I0129 12:08:56.454148 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9nh5h" podStartSLOduration=3.088125405 podStartE2EDuration="1m6.454141825s" podCreationTimestamp="2026-01-29 12:07:50 +0000 UTC" firstStartedPulling="2026-01-29 12:07:52.762872132 +0000 UTC m=+156.778002258" lastFinishedPulling="2026-01-29 12:08:56.128888552 +0000 UTC m=+220.144018678" observedRunningTime="2026-01-29 12:08:56.450634223 +0000 UTC m=+220.465764349" watchObservedRunningTime="2026-01-29 12:08:56.454141825 +0000 UTC m=+220.469271951" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.434024 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerStarted","Data":"561d45ed070a8ed87043292f64d3ff1180a8f2932cb84545f2687e9fa861d7a4"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.436669 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerStarted","Data":"6e73e049b248259e09d0889030137fc7c277543f8ab3dc1f32499e5301475c4c"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.439000 4993 generic.go:334] "Generic (PLEG): container finished" podID="af46a836-d45e-493b-9503-40281a2149d5" containerID="1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5" exitCode=0 Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.439070 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerDied","Data":"1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.442564 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerStarted","Data":"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.444756 4993 generic.go:334] "Generic (PLEG): container finished" podID="98f5acfe-223f-459b-a995-ae89a695ac32" containerID="7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6" exitCode=0 Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.444930 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerDied","Data":"7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.446633 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerStarted","Data":"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed"} Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.486653 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7lksb" podStartSLOduration=3.632387941 podStartE2EDuration="1m10.486636466s" podCreationTimestamp="2026-01-29 12:07:47 +0000 UTC" firstStartedPulling="2026-01-29 12:07:50.039469914 +0000 UTC m=+154.054600030" lastFinishedPulling="2026-01-29 12:08:56.893718429 +0000 UTC m=+220.908848555" observedRunningTime="2026-01-29 12:08:57.463227134 +0000 UTC m=+221.478357270" watchObservedRunningTime="2026-01-29 12:08:57.486636466 +0000 UTC m=+221.501766592" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.517224 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dlzds" podStartSLOduration=2.421863685 podStartE2EDuration="1m10.517205038s" podCreationTimestamp="2026-01-29 12:07:47 +0000 UTC" firstStartedPulling="2026-01-29 12:07:48.977122265 +0000 UTC m=+152.992252391" lastFinishedPulling="2026-01-29 12:08:57.072463618 +0000 UTC m=+221.087593744" observedRunningTime="2026-01-29 12:08:57.492746472 +0000 UTC m=+221.507876608" watchObservedRunningTime="2026-01-29 12:08:57.517205038 +0000 UTC m=+221.532335284" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.529287 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.529337 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.563825 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n7xrp" podStartSLOduration=1.764906093 podStartE2EDuration="1m8.563807454s" podCreationTimestamp="2026-01-29 12:07:49 +0000 UTC" firstStartedPulling="2026-01-29 12:07:50.056039743 +0000 UTC m=+154.071169879" lastFinishedPulling="2026-01-29 12:08:56.854941114 +0000 UTC m=+220.870071240" observedRunningTime="2026-01-29 12:08:57.540286519 +0000 UTC m=+221.555416655" watchObservedRunningTime="2026-01-29 12:08:57.563807454 +0000 UTC m=+221.578937600" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.569603 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p67p4" podStartSLOduration=3.3692548 podStartE2EDuration="1m8.569588059s" podCreationTimestamp="2026-01-29 12:07:49 +0000 UTC" firstStartedPulling="2026-01-29 12:07:51.85441385 +0000 UTC m=+155.869543976" lastFinishedPulling="2026-01-29 12:08:57.054747109 +0000 UTC m=+221.069877235" observedRunningTime="2026-01-29 12:08:57.563626898 +0000 UTC m=+221.578757034" watchObservedRunningTime="2026-01-29 12:08:57.569588059 +0000 UTC m=+221.584718175" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.605597 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.605667 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.875453 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:08:57 crc kubenswrapper[4993]: I0129 12:08:57.875729 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.456066 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerStarted","Data":"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4"} Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.458841 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerStarted","Data":"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b"} Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.485815 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-86rvn" podStartSLOduration=3.373274361 podStartE2EDuration="1m8.485796736s" podCreationTimestamp="2026-01-29 12:07:50 +0000 UTC" firstStartedPulling="2026-01-29 12:07:52.740451904 +0000 UTC m=+156.755582030" lastFinishedPulling="2026-01-29 12:08:57.852974279 +0000 UTC m=+221.868104405" observedRunningTime="2026-01-29 12:08:58.482549032 +0000 UTC m=+222.497679168" watchObservedRunningTime="2026-01-29 12:08:58.485796736 +0000 UTC m=+222.500926862" Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.505115 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zzw5p" podStartSLOduration=3.503298602 podStartE2EDuration="1m12.505095846s" podCreationTimestamp="2026-01-29 12:07:46 +0000 UTC" firstStartedPulling="2026-01-29 12:07:48.913445084 +0000 UTC m=+152.928575210" lastFinishedPulling="2026-01-29 12:08:57.915242328 +0000 UTC m=+221.930372454" observedRunningTime="2026-01-29 12:08:58.503889167 +0000 UTC m=+222.519019293" watchObservedRunningTime="2026-01-29 12:08:58.505095846 +0000 UTC m=+222.520225972" Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.756957 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dlzds" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="registry-server" probeResult="failure" output=< Jan 29 12:08:58 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:08:58 crc kubenswrapper[4993]: > Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.758102 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mfnw5" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="registry-server" probeResult="failure" output=< Jan 29 12:08:58 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:08:58 crc kubenswrapper[4993]: > Jan 29 12:08:58 crc kubenswrapper[4993]: I0129 12:08:58.927268 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7lksb" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="registry-server" probeResult="failure" output=< Jan 29 12:08:58 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:08:58 crc kubenswrapper[4993]: > Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.388763 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.388841 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.435224 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.813575 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.813942 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:08:59 crc kubenswrapper[4993]: I0129 12:08:59.854047 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:09:00 crc kubenswrapper[4993]: I0129 12:09:00.421840 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:09:00 crc kubenswrapper[4993]: I0129 12:09:00.421892 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:09:00 crc kubenswrapper[4993]: I0129 12:09:00.799218 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:09:00 crc kubenswrapper[4993]: I0129 12:09:00.799561 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:09:01 crc kubenswrapper[4993]: I0129 12:09:01.464961 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9nh5h" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="registry-server" probeResult="failure" output=< Jan 29 12:09:01 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:09:01 crc kubenswrapper[4993]: > Jan 29 12:09:01 crc kubenswrapper[4993]: I0129 12:09:01.840380 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-86rvn" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="registry-server" probeResult="failure" output=< Jan 29 12:09:01 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:09:01 crc kubenswrapper[4993]: > Jan 29 12:09:06 crc kubenswrapper[4993]: I0129 12:09:06.994765 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:09:06 crc kubenswrapper[4993]: I0129 12:09:06.995464 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.034976 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.554089 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.586427 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.636613 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.659814 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.694573 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.921969 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:09:07 crc kubenswrapper[4993]: I0129 12:09:07.968466 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:09:08 crc kubenswrapper[4993]: I0129 12:09:08.669394 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:09:09 crc kubenswrapper[4993]: I0129 12:09:09.428404 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:09:09 crc kubenswrapper[4993]: I0129 12:09:09.519781 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mfnw5" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="registry-server" containerID="cri-o://3904233c3a323ecbf0b9d65c3d54252e4299a85773b27eca90f1b6d12eaf4298" gracePeriod=2 Jan 29 12:09:09 crc kubenswrapper[4993]: I0129 12:09:09.867286 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.071527 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.071720 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7lksb" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="registry-server" containerID="cri-o://561d45ed070a8ed87043292f64d3ff1180a8f2932cb84545f2687e9fa861d7a4" gracePeriod=2 Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.461812 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.507002 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.853704 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:09:10 crc kubenswrapper[4993]: I0129 12:09:10.897068 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:09:11 crc kubenswrapper[4993]: I0129 12:09:11.533464 4993 generic.go:334] "Generic (PLEG): container finished" podID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerID="561d45ed070a8ed87043292f64d3ff1180a8f2932cb84545f2687e9fa861d7a4" exitCode=0 Jan 29 12:09:11 crc kubenswrapper[4993]: I0129 12:09:11.533524 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerDied","Data":"561d45ed070a8ed87043292f64d3ff1180a8f2932cb84545f2687e9fa861d7a4"} Jan 29 12:09:11 crc kubenswrapper[4993]: I0129 12:09:11.535580 4993 generic.go:334] "Generic (PLEG): container finished" podID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerID="3904233c3a323ecbf0b9d65c3d54252e4299a85773b27eca90f1b6d12eaf4298" exitCode=0 Jan 29 12:09:11 crc kubenswrapper[4993]: I0129 12:09:11.536295 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerDied","Data":"3904233c3a323ecbf0b9d65c3d54252e4299a85773b27eca90f1b6d12eaf4298"} Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.472516 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p67p4"] Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.473107 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p67p4" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="registry-server" containerID="cri-o://6e73e049b248259e09d0889030137fc7c277543f8ab3dc1f32499e5301475c4c" gracePeriod=2 Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.545872 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfnw5" event={"ID":"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb","Type":"ContainerDied","Data":"afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984"} Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.545921 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afd6721f4cf6e1bd146ed7f564eb4641036c7c1d2125be4d26630dea8e36b984" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.558156 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.607587 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psvt5\" (UniqueName: \"kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5\") pod \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.607761 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities\") pod \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.607832 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content\") pod \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\" (UID: \"f7c26ff5-7db3-40d9-bc94-56ee28d89dcb\") " Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.613610 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities" (OuterVolumeSpecName: "utilities") pod "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" (UID: "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.623397 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5" (OuterVolumeSpecName: "kube-api-access-psvt5") pod "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" (UID: "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb"). InnerVolumeSpecName "kube-api-access-psvt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.659244 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" (UID: "f7c26ff5-7db3-40d9-bc94-56ee28d89dcb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.709760 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.709810 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.709824 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psvt5\" (UniqueName: \"kubernetes.io/projected/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb-kube-api-access-psvt5\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:12 crc kubenswrapper[4993]: I0129 12:09:12.999178 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.114886 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content\") pod \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.115006 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities\") pod \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.115041 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm4tk\" (UniqueName: \"kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk\") pod \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\" (UID: \"ceb99f3c-f71e-4e45-9336-bac8338c94b7\") " Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.115835 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities" (OuterVolumeSpecName: "utilities") pod "ceb99f3c-f71e-4e45-9336-bac8338c94b7" (UID: "ceb99f3c-f71e-4e45-9336-bac8338c94b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.118321 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk" (OuterVolumeSpecName: "kube-api-access-rm4tk") pod "ceb99f3c-f71e-4e45-9336-bac8338c94b7" (UID: "ceb99f3c-f71e-4e45-9336-bac8338c94b7"). InnerVolumeSpecName "kube-api-access-rm4tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.162237 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ceb99f3c-f71e-4e45-9336-bac8338c94b7" (UID: "ceb99f3c-f71e-4e45-9336-bac8338c94b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.217009 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.217052 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb99f3c-f71e-4e45-9336-bac8338c94b7-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.217061 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm4tk\" (UniqueName: \"kubernetes.io/projected/ceb99f3c-f71e-4e45-9336-bac8338c94b7-kube-api-access-rm4tk\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.554104 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfnw5" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.554099 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lksb" event={"ID":"ceb99f3c-f71e-4e45-9336-bac8338c94b7","Type":"ContainerDied","Data":"a1c7b272cf2cee25115eb0b1645af5760721a83f95f298e6465f722c91a3dad2"} Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.554172 4993 scope.go:117] "RemoveContainer" containerID="561d45ed070a8ed87043292f64d3ff1180a8f2932cb84545f2687e9fa861d7a4" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.555510 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lksb" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.579428 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.586924 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mfnw5"] Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.592258 4993 scope.go:117] "RemoveContainer" containerID="eaba1b246b8f7dd6c086020e84b066146f338f16e8c6b7b27d75632fd43a6538" Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.594580 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.600898 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7lksb"] Jan 29 12:09:13 crc kubenswrapper[4993]: I0129 12:09:13.609480 4993 scope.go:117] "RemoveContainer" containerID="95d35d33904b9b205a0fc1fb618988c97df99fd87d1f405a82e6d840f8bca933" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.566833 4993 generic.go:334] "Generic (PLEG): container finished" podID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerID="6e73e049b248259e09d0889030137fc7c277543f8ab3dc1f32499e5301475c4c" exitCode=0 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.566916 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerDied","Data":"6e73e049b248259e09d0889030137fc7c277543f8ab3dc1f32499e5301475c4c"} Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.817752 4993 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818007 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="extract-content" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818020 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="extract-content" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818035 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="extract-content" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818042 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="extract-content" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818052 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="extract-utilities" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818058 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="extract-utilities" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818066 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="extract-utilities" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818072 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="extract-utilities" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818082 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818087 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818096 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818103 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.818117 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825bd7fd-af79-43b0-9aef-28a7ee9a2417" containerName="pruner" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818123 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="825bd7fd-af79-43b0-9aef-28a7ee9a2417" containerName="pruner" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818272 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="825bd7fd-af79-43b0-9aef-28a7ee9a2417" containerName="pruner" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818288 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.818297 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" containerName="registry-server" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.819079 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.819214 4993 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.820912 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a" gracePeriod=15 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.821350 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4" gracePeriod=15 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.821499 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04" gracePeriod=15 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.821445 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d" gracePeriod=15 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.821807 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e" gracePeriod=15 Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.826821 4993 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827532 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827554 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827570 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827578 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827590 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827600 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827615 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827623 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827634 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827641 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827657 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827664 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827680 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827687 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.827706 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.827714 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829481 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829545 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829580 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829598 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829626 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.829643 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.830387 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 12:09:14 crc kubenswrapper[4993]: E0129 12:09:14.860254 4993 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941051 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941333 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941435 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941518 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941600 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941685 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941762 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:14 crc kubenswrapper[4993]: I0129 12:09:14.941893 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.043890 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.043933 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.043979 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.043999 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044017 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044037 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044058 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044071 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044130 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044163 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044197 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044227 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044247 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044266 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044286 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.044306 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.160847 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.160968 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.161883 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.162093 4993 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: W0129 12:09:15.178397 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-fd9bcf40d3ac095a65be5a7f1a1a57ba8f26caddb8e76c1c5cd38956ed779ed0 WatchSource:0}: Error finding container fd9bcf40d3ac095a65be5a7f1a1a57ba8f26caddb8e76c1c5cd38956ed779ed0: Status 404 returned error can't find the container with id fd9bcf40d3ac095a65be5a7f1a1a57ba8f26caddb8e76c1c5cd38956ed779ed0 Jan 29 12:09:15 crc kubenswrapper[4993]: E0129 12:09:15.182200 4993 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f325d5eaca0b3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,LastTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.198359 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb99f3c-f71e-4e45-9336-bac8338c94b7" path="/var/lib/kubelet/pods/ceb99f3c-f71e-4e45-9336-bac8338c94b7/volumes" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.199076 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c26ff5-7db3-40d9-bc94-56ee28d89dcb" path="/var/lib/kubelet/pods/f7c26ff5-7db3-40d9-bc94-56ee28d89dcb/volumes" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.246106 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjvrz\" (UniqueName: \"kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz\") pod \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.246167 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content\") pod \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.246231 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities\") pod \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\" (UID: \"8421bbe0-da5e-49ec-994f-fd9a072a4bc7\") " Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.247145 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities" (OuterVolumeSpecName: "utilities") pod "8421bbe0-da5e-49ec-994f-fd9a072a4bc7" (UID: "8421bbe0-da5e-49ec-994f-fd9a072a4bc7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.248932 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz" (OuterVolumeSpecName: "kube-api-access-hjvrz") pod "8421bbe0-da5e-49ec-994f-fd9a072a4bc7" (UID: "8421bbe0-da5e-49ec-994f-fd9a072a4bc7"). InnerVolumeSpecName "kube-api-access-hjvrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.270495 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8421bbe0-da5e-49ec-994f-fd9a072a4bc7" (UID: "8421bbe0-da5e-49ec-994f-fd9a072a4bc7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.347673 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjvrz\" (UniqueName: \"kubernetes.io/projected/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-kube-api-access-hjvrz\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.347703 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.348515 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8421bbe0-da5e-49ec-994f-fd9a072a4bc7-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.575521 4993 generic.go:334] "Generic (PLEG): container finished" podID="39088c04-e684-40fc-bdec-8933958c2a4e" containerID="c3af66244ef21d53950a1f1d1867e20c93215d02e78feee392ac0d69f38f0c73" exitCode=0 Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.575680 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"39088c04-e684-40fc-bdec-8933958c2a4e","Type":"ContainerDied","Data":"c3af66244ef21d53950a1f1d1867e20c93215d02e78feee392ac0d69f38f0c73"} Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.576633 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.576987 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.578287 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31"} Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.578323 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"fd9bcf40d3ac095a65be5a7f1a1a57ba8f26caddb8e76c1c5cd38956ed779ed0"} Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.580506 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.582118 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.583763 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e" exitCode=0 Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.583789 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4" exitCode=0 Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.583802 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d" exitCode=0 Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.583814 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04" exitCode=2 Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.583871 4993 scope.go:117] "RemoveContainer" containerID="0f25bd996045f115071c3d060a4693352d8b1a99a54c0764231396bd4b227bc0" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.587214 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p67p4" event={"ID":"8421bbe0-da5e-49ec-994f-fd9a072a4bc7","Type":"ContainerDied","Data":"874b0c2523b59170414f12b384e255f4bf2d60461f3450b44c5f99517a87403f"} Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.587313 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p67p4" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.588142 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.588426 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.605774 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.606096 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.650897 4993 scope.go:117] "RemoveContainer" containerID="6e73e049b248259e09d0889030137fc7c277543f8ab3dc1f32499e5301475c4c" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.670291 4993 scope.go:117] "RemoveContainer" containerID="6bb1c547ebc5847bf94dba64b05b4b6c5eca4d23254cc736c1890253ade50238" Jan 29 12:09:15 crc kubenswrapper[4993]: I0129 12:09:15.695917 4993 scope.go:117] "RemoveContainer" containerID="b16b2fdd35435f6dfb078fb999f07ffd5844d3b82f291cfc0fc82592d4c3f6f4" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.597907 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 12:09:16 crc kubenswrapper[4993]: E0129 12:09:16.600334 4993 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.600356 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.600673 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.824075 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.825178 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.825608 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.972818 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access\") pod \"39088c04-e684-40fc-bdec-8933958c2a4e\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.972912 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock\") pod \"39088c04-e684-40fc-bdec-8933958c2a4e\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.972959 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir\") pod \"39088c04-e684-40fc-bdec-8933958c2a4e\" (UID: \"39088c04-e684-40fc-bdec-8933958c2a4e\") " Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.973030 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock" (OuterVolumeSpecName: "var-lock") pod "39088c04-e684-40fc-bdec-8933958c2a4e" (UID: "39088c04-e684-40fc-bdec-8933958c2a4e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.973178 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "39088c04-e684-40fc-bdec-8933958c2a4e" (UID: "39088c04-e684-40fc-bdec-8933958c2a4e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.973215 4993 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:16 crc kubenswrapper[4993]: I0129 12:09:16.985514 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "39088c04-e684-40fc-bdec-8933958c2a4e" (UID: "39088c04-e684-40fc-bdec-8933958c2a4e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.074588 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/39088c04-e684-40fc-bdec-8933958c2a4e-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.074622 4993 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/39088c04-e684-40fc-bdec-8933958c2a4e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.193035 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.193308 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.194109 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.197257 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.198430 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.198774 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276435 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276528 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276544 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276589 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276654 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276730 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276857 4993 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276870 4993 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.276878 4993 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.608029 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.608086 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"39088c04-e684-40fc-bdec-8933958c2a4e","Type":"ContainerDied","Data":"9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a"} Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.608125 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e8625e5cca256bf62d42e83d435755cf6112742cb6b7566f98f43ac1e19a88a" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.613441 4993 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a" exitCode=0 Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.613519 4993 scope.go:117] "RemoveContainer" containerID="05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.613526 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.614350 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.614727 4993 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.614930 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.615276 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.615470 4993 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.615668 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.629123 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.629353 4993 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.629540 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.635496 4993 scope.go:117] "RemoveContainer" containerID="a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.654885 4993 scope.go:117] "RemoveContainer" containerID="3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.668055 4993 scope.go:117] "RemoveContainer" containerID="615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.689212 4993 scope.go:117] "RemoveContainer" containerID="51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.707961 4993 scope.go:117] "RemoveContainer" containerID="7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.725454 4993 scope.go:117] "RemoveContainer" containerID="05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.726045 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\": container with ID starting with 05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e not found: ID does not exist" containerID="05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726086 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e"} err="failed to get container status \"05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\": rpc error: code = NotFound desc = could not find container \"05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e\": container with ID starting with 05b5306b591b994619acd1f058e7f33d13bc6292814e104ff605322e8aad0f5e not found: ID does not exist" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726117 4993 scope.go:117] "RemoveContainer" containerID="a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.726516 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\": container with ID starting with a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4 not found: ID does not exist" containerID="a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726546 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4"} err="failed to get container status \"a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\": rpc error: code = NotFound desc = could not find container \"a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4\": container with ID starting with a902ce94355fb44c55c77d91ffa9478a9e60422d780400a5a61405efdd038cd4 not found: ID does not exist" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726559 4993 scope.go:117] "RemoveContainer" containerID="3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.726801 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\": container with ID starting with 3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d not found: ID does not exist" containerID="3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726825 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d"} err="failed to get container status \"3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\": rpc error: code = NotFound desc = could not find container \"3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d\": container with ID starting with 3c87c8c7a37c5eb2a9d5459d4268ccbeca98d8319498c6a4c5cbc077d833df1d not found: ID does not exist" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.726837 4993 scope.go:117] "RemoveContainer" containerID="615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.727171 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\": container with ID starting with 615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04 not found: ID does not exist" containerID="615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.727211 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04"} err="failed to get container status \"615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\": rpc error: code = NotFound desc = could not find container \"615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04\": container with ID starting with 615d5ca76a49223bf3d13cece318a56bc561d0f150336d17d70dd02c01d99b04 not found: ID does not exist" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.727228 4993 scope.go:117] "RemoveContainer" containerID="51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.728678 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\": container with ID starting with 51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a not found: ID does not exist" containerID="51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.728713 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a"} err="failed to get container status \"51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\": rpc error: code = NotFound desc = could not find container \"51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a\": container with ID starting with 51b936a5f5189dfde1cc942c59714dac96e86fce53e4a30415169da1cc2dd60a not found: ID does not exist" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.728734 4993 scope.go:117] "RemoveContainer" containerID="7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa" Jan 29 12:09:17 crc kubenswrapper[4993]: E0129 12:09:17.729390 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\": container with ID starting with 7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa not found: ID does not exist" containerID="7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa" Jan 29 12:09:17 crc kubenswrapper[4993]: I0129 12:09:17.729440 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa"} err="failed to get container status \"7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\": rpc error: code = NotFound desc = could not find container \"7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa\": container with ID starting with 7e91fb7d3114d2a7117151985486f4aa2b87ed406f78a4a17d745dd2a33455aa not found: ID does not exist" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.031055 4993 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f325d5eaca0b3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,LastTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.186370 4993 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.186910 4993 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.187343 4993 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.187543 4993 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.187821 4993 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: I0129 12:09:18.187891 4993 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.188323 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.390842 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723010 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:09:18Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:09:18Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:09:18Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T12:09:18Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723401 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723639 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723794 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723925 4993 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.723939 4993 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 12:09:18 crc kubenswrapper[4993]: E0129 12:09:18.792065 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Jan 29 12:09:19 crc kubenswrapper[4993]: I0129 12:09:19.201253 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 29 12:09:19 crc kubenswrapper[4993]: E0129 12:09:19.592799 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Jan 29 12:09:21 crc kubenswrapper[4993]: E0129 12:09:21.193426 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Jan 29 12:09:24 crc kubenswrapper[4993]: E0129 12:09:24.394845 4993 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="6.4s" Jan 29 12:09:27 crc kubenswrapper[4993]: I0129 12:09:27.193159 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:27 crc kubenswrapper[4993]: I0129 12:09:27.194160 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:28 crc kubenswrapper[4993]: E0129 12:09:28.032954 4993 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f325d5eaca0b3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,LastTimestamp:2026-01-29 12:09:15.181613235 +0000 UTC m=+239.196743361,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.677512 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.677571 4993 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f" exitCode=1 Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.677608 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f"} Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.678145 4993 scope.go:117] "RemoveContainer" containerID="5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f" Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.678551 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.679045 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:28 crc kubenswrapper[4993]: I0129 12:09:28.679613 4993 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.189808 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.191343 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.191986 4993 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.192579 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.209808 4993 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.209862 4993 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:29 crc kubenswrapper[4993]: E0129 12:09:29.210501 4993 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.211017 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:29 crc kubenswrapper[4993]: W0129 12:09:29.230689 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b8406edfca47866c9742022f9635a60eac01c82af4ad57db7c9490ce74e41be6 WatchSource:0}: Error finding container b8406edfca47866c9742022f9635a60eac01c82af4ad57db7c9490ce74e41be6: Status 404 returned error can't find the container with id b8406edfca47866c9742022f9635a60eac01c82af4ad57db7c9490ce74e41be6 Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.691443 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.691722 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f38e4d7de1637b445fb1d69483c28aef9ea4d195d817fe1af186450ef4cc737d"} Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.692969 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.693528 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.694016 4993 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.695279 4993 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9ca79af46266d61586192ca1fa6bf02cc6e238b30af6e64d74759f5029c4586d" exitCode=0 Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.695336 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9ca79af46266d61586192ca1fa6bf02cc6e238b30af6e64d74759f5029c4586d"} Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.695415 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b8406edfca47866c9742022f9635a60eac01c82af4ad57db7c9490ce74e41be6"} Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.695839 4993 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.695871 4993 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.696278 4993 status_manager.go:851] "Failed to get status for pod" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: E0129 12:09:29.696359 4993 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.696594 4993 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:29 crc kubenswrapper[4993]: I0129 12:09:29.697075 4993 status_manager.go:851] "Failed to get status for pod" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" pod="openshift-marketplace/redhat-marketplace-p67p4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-p67p4\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 12:09:30 crc kubenswrapper[4993]: I0129 12:09:30.704470 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8a4849e715405e7eab94dde897e555c4807fc651e490d3a3f213232bdf839e56"} Jan 29 12:09:30 crc kubenswrapper[4993]: I0129 12:09:30.704840 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"351d2213dc72d7a418aa0d4eaa5bb1d85838388b64828cf67a9506e7cc23dc10"} Jan 29 12:09:30 crc kubenswrapper[4993]: I0129 12:09:30.704855 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f63f2c8a63ac5a71900d17a218aca9b0e855e805cd6af36ee5dc7cb1634d1cb0"} Jan 29 12:09:30 crc kubenswrapper[4993]: I0129 12:09:30.704867 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f38721a2651fa5ae8acbbcbd5f74dc695d390eb00bd4120b6e790098e7a33a8a"} Jan 29 12:09:31 crc kubenswrapper[4993]: I0129 12:09:31.714842 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1cf6892d8814e857d4d6179a82b0aea93a29446616dd0d3fd95dba1d38f4b624"} Jan 29 12:09:31 crc kubenswrapper[4993]: I0129 12:09:31.715038 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:31 crc kubenswrapper[4993]: I0129 12:09:31.715150 4993 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:31 crc kubenswrapper[4993]: I0129 12:09:31.715176 4993 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.212100 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.213228 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.219001 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.835757 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.836008 4993 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 12:09:34 crc kubenswrapper[4993]: I0129 12:09:34.836112 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 12:09:36 crc kubenswrapper[4993]: I0129 12:09:36.725902 4993 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.210992 4993 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7fd165ec-8992-45ae-8e8d-3937b64f89d4" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.744022 4993 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.744075 4993 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.747330 4993 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7fd165ec-8992-45ae-8e8d-3937b64f89d4" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.747951 4993 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://f38721a2651fa5ae8acbbcbd5f74dc695d390eb00bd4120b6e790098e7a33a8a" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.747976 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:37 crc kubenswrapper[4993]: I0129 12:09:37.757955 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:09:38 crc kubenswrapper[4993]: I0129 12:09:38.748366 4993 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:38 crc kubenswrapper[4993]: I0129 12:09:38.748394 4993 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="84322ce4-8377-423b-8c5d-cf2dd90ad0e1" Jan 29 12:09:38 crc kubenswrapper[4993]: I0129 12:09:38.752167 4993 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7fd165ec-8992-45ae-8e8d-3937b64f89d4" Jan 29 12:09:44 crc kubenswrapper[4993]: I0129 12:09:44.836164 4993 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 12:09:44 crc kubenswrapper[4993]: I0129 12:09:44.836822 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 12:09:45 crc kubenswrapper[4993]: I0129 12:09:45.729318 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 12:09:46 crc kubenswrapper[4993]: I0129 12:09:46.848602 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 12:09:46 crc kubenswrapper[4993]: I0129 12:09:46.894450 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 12:09:46 crc kubenswrapper[4993]: I0129 12:09:46.938997 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.059167 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.381438 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.405465 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.441099 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.447750 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.972834 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 12:09:47 crc kubenswrapper[4993]: I0129 12:09:47.997664 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.010581 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.027743 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.217961 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.251101 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.504801 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.507029 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.612237 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.615679 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.628257 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.685141 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.786693 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.884117 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.972358 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 12:09:48 crc kubenswrapper[4993]: I0129 12:09:48.994146 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.095753 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.150537 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.160918 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.281093 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.342882 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.353723 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.397116 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.495557 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.495772 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.507064 4993 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.512348 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p67p4","openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.512431 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.517915 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.532643 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.532629209 podStartE2EDuration="13.532629209s" podCreationTimestamp="2026-01-29 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:09:49.528866671 +0000 UTC m=+273.543996807" watchObservedRunningTime="2026-01-29 12:09:49.532629209 +0000 UTC m=+273.547759335" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.628013 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 12:09:49 crc kubenswrapper[4993]: I0129 12:09:49.672838 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.056150 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.138476 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.154151 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.205710 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.264182 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.297342 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.302764 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.353123 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.421470 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.432485 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.469634 4993 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.470901 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.491458 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.494114 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.584056 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.604100 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.626466 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.784704 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.786758 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.829530 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.931593 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 12:09:50 crc kubenswrapper[4993]: I0129 12:09:50.959940 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.112373 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.166268 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.197598 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" path="/var/lib/kubelet/pods/8421bbe0-da5e-49ec-994f-fd9a072a4bc7/volumes" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.198531 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.204809 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.254026 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.312601 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.380413 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.423030 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.466641 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.473337 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.557248 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.568348 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.580568 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.583269 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.593669 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.657611 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.667062 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.699835 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.757493 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.830357 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.876166 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.878291 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.946830 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 12:09:51 crc kubenswrapper[4993]: I0129 12:09:51.990448 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.040807 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.077980 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.129002 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.134777 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.254197 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.295773 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.328392 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.361389 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.364594 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.396656 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.427034 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.491915 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.539426 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.557310 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.566216 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.811731 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 12:09:52 crc kubenswrapper[4993]: I0129 12:09:52.997121 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.006623 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.124323 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.187090 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.243971 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.246595 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.310385 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.324117 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.325413 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.385612 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.424654 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.449668 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.498033 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.670421 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.680079 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.784180 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.796699 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.820473 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.845682 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.950618 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 12:09:53 crc kubenswrapper[4993]: I0129 12:09:53.954959 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.089837 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.151251 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.159132 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.295026 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.327969 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.363037 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.405469 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.456411 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.492809 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.615579 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.632224 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.635532 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.782652 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.807043 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.836125 4993 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.836233 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.836284 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.836893 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"f38e4d7de1637b445fb1d69483c28aef9ea4d195d817fe1af186450ef4cc737d"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.837008 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://f38e4d7de1637b445fb1d69483c28aef9ea4d195d817fe1af186450ef4cc737d" gracePeriod=30 Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.880027 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.901506 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.917770 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 12:09:54 crc kubenswrapper[4993]: I0129 12:09:54.931516 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.112089 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.114178 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.133411 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.155958 4993 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.213021 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.221900 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.539224 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.553559 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.599367 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.640733 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.672715 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.841106 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.879626 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.884061 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.908777 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.947675 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.973635 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 12:09:55 crc kubenswrapper[4993]: I0129 12:09:55.988747 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.007387 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.062981 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.072335 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.101150 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.108720 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.206291 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.207020 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.257427 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.333987 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.454434 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.468743 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.557368 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.660003 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.743137 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.804986 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.845150 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.913451 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 12:09:56 crc kubenswrapper[4993]: I0129 12:09:56.959919 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.077773 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.159327 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.171530 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.181315 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.262275 4993 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.284288 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.341668 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.365793 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.539750 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.612727 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.645904 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.717481 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.720586 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.741427 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.808829 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.808852 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.877038 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.912884 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.989032 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 12:09:57 crc kubenswrapper[4993]: I0129 12:09:57.994549 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.154691 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.227396 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.426272 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.430009 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.462679 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.480612 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.495572 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.650618 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.702694 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.806264 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.878358 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.878947 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.893274 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.932131 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.954043 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.964454 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.974848 4993 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.975402 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31" gracePeriod=5 Jan 29 12:09:58 crc kubenswrapper[4993]: I0129 12:09:58.978726 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.186737 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.200535 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.219033 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.441421 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.503744 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.536442 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.572169 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.616141 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.623407 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.779457 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 12:09:59 crc kubenswrapper[4993]: I0129 12:09:59.794132 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.057337 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.130497 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.170273 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.240450 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.329520 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.348976 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.394163 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.400940 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.498717 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.594539 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.625903 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.781706 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.806001 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.869085 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 12:10:00 crc kubenswrapper[4993]: I0129 12:10:00.934147 4993 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.031848 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.121708 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.273102 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.311421 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.366950 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.664764 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.769060 4993 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.799151 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.806618 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.852658 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 12:10:01 crc kubenswrapper[4993]: I0129 12:10:01.983554 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 12:10:02 crc kubenswrapper[4993]: I0129 12:10:02.054535 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 12:10:02 crc kubenswrapper[4993]: I0129 12:10:02.099777 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 12:10:02 crc kubenswrapper[4993]: I0129 12:10:02.488246 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.452314 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.572788 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.572926 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683280 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683333 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683424 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683468 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683499 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683548 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683656 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683680 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.683747 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.684094 4993 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.684133 4993 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.684152 4993 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.684169 4993 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.695605 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.784812 4993 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.895913 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.895960 4993 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31" exitCode=137 Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.896001 4993 scope.go:117] "RemoveContainer" containerID="47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.896100 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.916769 4993 scope.go:117] "RemoveContainer" containerID="47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31" Jan 29 12:10:04 crc kubenswrapper[4993]: E0129 12:10:04.917164 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31\": container with ID starting with 47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31 not found: ID does not exist" containerID="47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31" Jan 29 12:10:04 crc kubenswrapper[4993]: I0129 12:10:04.917244 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31"} err="failed to get container status \"47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31\": rpc error: code = NotFound desc = could not find container \"47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31\": container with ID starting with 47d5d088eb34fe25803efe0c30fe46dd216f056ebd877624db9a7f576f1def31 not found: ID does not exist" Jan 29 12:10:05 crc kubenswrapper[4993]: I0129 12:10:05.199985 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 29 12:10:15 crc kubenswrapper[4993]: I0129 12:10:15.962638 4993 generic.go:334] "Generic (PLEG): container finished" podID="41f046c9-8758-4f06-a7d4-8149883734eb" containerID="139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904" exitCode=0 Jan 29 12:10:15 crc kubenswrapper[4993]: I0129 12:10:15.962756 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerDied","Data":"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904"} Jan 29 12:10:15 crc kubenswrapper[4993]: I0129 12:10:15.963367 4993 scope.go:117] "RemoveContainer" containerID="139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904" Jan 29 12:10:16 crc kubenswrapper[4993]: I0129 12:10:16.968682 4993 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 29 12:10:16 crc kubenswrapper[4993]: I0129 12:10:16.969831 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerStarted","Data":"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f"} Jan 29 12:10:16 crc kubenswrapper[4993]: I0129 12:10:16.970221 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:10:16 crc kubenswrapper[4993]: I0129 12:10:16.972350 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.020020 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.022320 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.022384 4993 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f38e4d7de1637b445fb1d69483c28aef9ea4d195d817fe1af186450ef4cc737d" exitCode=137 Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.022428 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f38e4d7de1637b445fb1d69483c28aef9ea4d195d817fe1af186450ef4cc737d"} Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.022468 4993 scope.go:117] "RemoveContainer" containerID="5689d8275f799a965df4e819bb8f6b4422f81090b6166e7ec6fb97ac0ca98b5f" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.045828 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.046059 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-86rvn" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="registry-server" containerID="cri-o://4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4" gracePeriod=2 Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.475095 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.555300 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities\") pod \"af46a836-d45e-493b-9503-40281a2149d5\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.555413 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content\") pod \"af46a836-d45e-493b-9503-40281a2149d5\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.555513 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkgjf\" (UniqueName: \"kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf\") pod \"af46a836-d45e-493b-9503-40281a2149d5\" (UID: \"af46a836-d45e-493b-9503-40281a2149d5\") " Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.556653 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities" (OuterVolumeSpecName: "utilities") pod "af46a836-d45e-493b-9503-40281a2149d5" (UID: "af46a836-d45e-493b-9503-40281a2149d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.561170 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf" (OuterVolumeSpecName: "kube-api-access-qkgjf") pod "af46a836-d45e-493b-9503-40281a2149d5" (UID: "af46a836-d45e-493b-9503-40281a2149d5"). InnerVolumeSpecName "kube-api-access-qkgjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.656519 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.656575 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkgjf\" (UniqueName: \"kubernetes.io/projected/af46a836-d45e-493b-9503-40281a2149d5-kube-api-access-qkgjf\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.679899 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af46a836-d45e-493b-9503-40281a2149d5" (UID: "af46a836-d45e-493b-9503-40281a2149d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:10:25 crc kubenswrapper[4993]: I0129 12:10:25.758019 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af46a836-d45e-493b-9503-40281a2149d5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.030616 4993 generic.go:334] "Generic (PLEG): container finished" podID="af46a836-d45e-493b-9503-40281a2149d5" containerID="4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4" exitCode=0 Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.030663 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-86rvn" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.030680 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerDied","Data":"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4"} Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.031238 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-86rvn" event={"ID":"af46a836-d45e-493b-9503-40281a2149d5","Type":"ContainerDied","Data":"a1e516b6e0706fb6b42a33c12b7ac926b650d9694f29579ca5a706c7bed059a6"} Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.031285 4993 scope.go:117] "RemoveContainer" containerID="4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.034604 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.035976 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a91ab30ce45f408e48643d6ff89552f01f9abdec6fc68e1506ad2e96ade44096"} Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.045001 4993 scope.go:117] "RemoveContainer" containerID="1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.070548 4993 scope.go:117] "RemoveContainer" containerID="855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.070758 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.074034 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-86rvn"] Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.117525 4993 scope.go:117] "RemoveContainer" containerID="4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4" Jan 29 12:10:26 crc kubenswrapper[4993]: E0129 12:10:26.118560 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4\": container with ID starting with 4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4 not found: ID does not exist" containerID="4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.118787 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4"} err="failed to get container status \"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4\": rpc error: code = NotFound desc = could not find container \"4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4\": container with ID starting with 4553cbbb14862356b86ae1828b234b571686d472bc224a3c1238cce3f99946a4 not found: ID does not exist" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.118983 4993 scope.go:117] "RemoveContainer" containerID="1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5" Jan 29 12:10:26 crc kubenswrapper[4993]: E0129 12:10:26.119694 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5\": container with ID starting with 1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5 not found: ID does not exist" containerID="1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.119742 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5"} err="failed to get container status \"1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5\": rpc error: code = NotFound desc = could not find container \"1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5\": container with ID starting with 1424f39d77524a1f2369ed451556d574992f98c8ccbd41db8d7fef6e9c9ed2a5 not found: ID does not exist" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.119775 4993 scope.go:117] "RemoveContainer" containerID="855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c" Jan 29 12:10:26 crc kubenswrapper[4993]: E0129 12:10:26.121164 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c\": container with ID starting with 855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c not found: ID does not exist" containerID="855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c" Jan 29 12:10:26 crc kubenswrapper[4993]: I0129 12:10:26.121205 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c"} err="failed to get container status \"855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c\": rpc error: code = NotFound desc = could not find container \"855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c\": container with ID starting with 855d4675e5c385539eaa8df6920d54598c8eb1bb5cec5d53849973bb6b97d43c not found: ID does not exist" Jan 29 12:10:27 crc kubenswrapper[4993]: I0129 12:10:27.197565 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af46a836-d45e-493b-9503-40281a2149d5" path="/var/lib/kubelet/pods/af46a836-d45e-493b-9503-40281a2149d5/volumes" Jan 29 12:10:27 crc kubenswrapper[4993]: I0129 12:10:27.758422 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:10:34 crc kubenswrapper[4993]: I0129 12:10:34.836034 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:10:34 crc kubenswrapper[4993]: I0129 12:10:34.839269 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:10:35 crc kubenswrapper[4993]: I0129 12:10:35.087912 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 12:10:45 crc kubenswrapper[4993]: I0129 12:10:45.500300 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:10:45 crc kubenswrapper[4993]: I0129 12:10:45.501154 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" podUID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" containerName="route-controller-manager" containerID="cri-o://5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f" gracePeriod=30 Jan 29 12:10:45 crc kubenswrapper[4993]: I0129 12:10:45.504707 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:10:45 crc kubenswrapper[4993]: I0129 12:10:45.504964 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" podUID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" containerName="controller-manager" containerID="cri-o://eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01" gracePeriod=30 Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.032462 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.036518 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.113889 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca\") pod \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.113934 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca\") pod \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.113981 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vznjm\" (UniqueName: \"kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm\") pod \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114027 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert\") pod \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114044 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert\") pod \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114073 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config\") pod \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114089 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv7lv\" (UniqueName: \"kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv\") pod \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114116 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config\") pod \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\" (UID: \"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114133 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles\") pod \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\" (UID: \"4bfb01f8-c2bf-4975-9fdc-bbf303c00295\") " Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114862 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" (UID: "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114892 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4bfb01f8-c2bf-4975-9fdc-bbf303c00295" (UID: "4bfb01f8-c2bf-4975-9fdc-bbf303c00295"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114933 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca" (OuterVolumeSpecName: "client-ca") pod "4bfb01f8-c2bf-4975-9fdc-bbf303c00295" (UID: "4bfb01f8-c2bf-4975-9fdc-bbf303c00295"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.114988 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config" (OuterVolumeSpecName: "config") pod "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" (UID: "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.115748 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config" (OuterVolumeSpecName: "config") pod "4bfb01f8-c2bf-4975-9fdc-bbf303c00295" (UID: "4bfb01f8-c2bf-4975-9fdc-bbf303c00295"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.119540 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv" (OuterVolumeSpecName: "kube-api-access-hv7lv") pod "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" (UID: "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6"). InnerVolumeSpecName "kube-api-access-hv7lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.121064 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" (UID: "5ec74443-272b-4f3e-b4f8-8cf5c62db9a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.121539 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4bfb01f8-c2bf-4975-9fdc-bbf303c00295" (UID: "4bfb01f8-c2bf-4975-9fdc-bbf303c00295"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.122807 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm" (OuterVolumeSpecName: "kube-api-access-vznjm") pod "4bfb01f8-c2bf-4975-9fdc-bbf303c00295" (UID: "4bfb01f8-c2bf-4975-9fdc-bbf303c00295"). InnerVolumeSpecName "kube-api-access-vznjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.149000 4993 generic.go:334] "Generic (PLEG): container finished" podID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" containerID="5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f" exitCode=0 Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.149083 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" event={"ID":"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6","Type":"ContainerDied","Data":"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f"} Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.149119 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" event={"ID":"5ec74443-272b-4f3e-b4f8-8cf5c62db9a6","Type":"ContainerDied","Data":"bab02233140d9afa82f2964e716ee0a83e49665e155a9e5c3566ff9156aba45e"} Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.149144 4993 scope.go:117] "RemoveContainer" containerID="5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.149307 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.156854 4993 generic.go:334] "Generic (PLEG): container finished" podID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" containerID="eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01" exitCode=0 Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.156923 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.156943 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" event={"ID":"4bfb01f8-c2bf-4975-9fdc-bbf303c00295","Type":"ContainerDied","Data":"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01"} Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.157136 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d" event={"ID":"4bfb01f8-c2bf-4975-9fdc-bbf303c00295","Type":"ContainerDied","Data":"2e298a73fb6a7916d73cbff7f622449b3b723448609776381205af5b307a4838"} Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.187525 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.188761 4993 scope.go:117] "RemoveContainer" containerID="5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f" Jan 29 12:10:46 crc kubenswrapper[4993]: E0129 12:10:46.189276 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f\": container with ID starting with 5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f not found: ID does not exist" containerID="5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.189302 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f"} err="failed to get container status \"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f\": rpc error: code = NotFound desc = could not find container \"5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f\": container with ID starting with 5b1fd227cdb395a45559449b927fd6b1a5cbbd9f36ca2043ff538df89b86e19f not found: ID does not exist" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.189320 4993 scope.go:117] "RemoveContainer" containerID="eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.195572 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6bfb7f7847-cjs2d"] Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.209871 4993 scope.go:117] "RemoveContainer" containerID="eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01" Jan 29 12:10:46 crc kubenswrapper[4993]: E0129 12:10:46.210331 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01\": container with ID starting with eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01 not found: ID does not exist" containerID="eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.210361 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01"} err="failed to get container status \"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01\": rpc error: code = NotFound desc = could not find container \"eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01\": container with ID starting with eebcd24662b03674439d493661b25822990351b5e0655a2bfb2b54df1b489b01 not found: ID does not exist" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.212237 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214842 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vznjm\" (UniqueName: \"kubernetes.io/projected/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-kube-api-access-vznjm\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214868 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214879 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214889 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214897 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv7lv\" (UniqueName: \"kubernetes.io/projected/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-kube-api-access-hv7lv\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214905 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214913 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214920 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bfb01f8-c2bf-4975-9fdc-bbf303c00295-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.214927 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:10:46 crc kubenswrapper[4993]: I0129 12:10:46.216602 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b77dc47b6-vr2td"] Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.218345 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" path="/var/lib/kubelet/pods/4bfb01f8-c2bf-4975-9fdc-bbf303c00295/volumes" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.218888 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" path="/var/lib/kubelet/pods/5ec74443-272b-4f3e-b4f8-8cf5c62db9a6/volumes" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745381 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745681 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" containerName="route-controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745697 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" containerName="route-controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745711 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="extract-content" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745720 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="extract-content" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745733 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745741 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745762 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="extract-content" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745772 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="extract-content" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745783 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745790 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745804 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745812 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745821 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="extract-utilities" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745828 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="extract-utilities" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745838 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" containerName="controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745845 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" containerName="controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745856 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" containerName="installer" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745863 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" containerName="installer" Jan 29 12:10:47 crc kubenswrapper[4993]: E0129 12:10:47.745874 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="extract-utilities" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745881 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="extract-utilities" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745984 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.745998 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bfb01f8-c2bf-4975-9fdc-bbf303c00295" containerName="controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.746009 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec74443-272b-4f3e-b4f8-8cf5c62db9a6" containerName="route-controller-manager" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.746020 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="af46a836-d45e-493b-9503-40281a2149d5" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.746028 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="39088c04-e684-40fc-bdec-8933958c2a4e" containerName="installer" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.746039 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="8421bbe0-da5e-49ec-994f-fd9a072a4bc7" containerName="registry-server" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.746491 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.748221 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.749556 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.750282 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.750332 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.751679 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.754466 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.754655 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.754729 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.754800 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.755106 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.755337 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.755367 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.755455 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.755587 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.759608 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.761806 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.764112 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830109 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830374 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830405 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830422 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830444 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830574 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830695 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rp4t\" (UniqueName: \"kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830727 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvwgb\" (UniqueName: \"kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.830746 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932546 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932683 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rp4t\" (UniqueName: \"kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932729 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvwgb\" (UniqueName: \"kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932765 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932835 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932871 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932913 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932950 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.932988 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.934057 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.934095 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.934779 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.934812 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.935433 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.938698 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.939163 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.951279 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rp4t\" (UniqueName: \"kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t\") pod \"controller-manager-75795fb4fc-c2vxg\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:47 crc kubenswrapper[4993]: I0129 12:10:47.959587 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvwgb\" (UniqueName: \"kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb\") pod \"route-controller-manager-54769cd8f6-l8mmb\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:48 crc kubenswrapper[4993]: I0129 12:10:48.066520 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:48 crc kubenswrapper[4993]: I0129 12:10:48.079055 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:48 crc kubenswrapper[4993]: I0129 12:10:48.291533 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:10:48 crc kubenswrapper[4993]: I0129 12:10:48.319427 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.179442 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" event={"ID":"4d046304-d1f8-41e2-a8c0-885fe1aeefbd","Type":"ContainerStarted","Data":"d65b919e1b133b197030e38e5835e68e7d95621cc8c162a9b9259e09b0da5f53"} Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.180030 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.180047 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" event={"ID":"4d046304-d1f8-41e2-a8c0-885fe1aeefbd","Type":"ContainerStarted","Data":"6f5d9fa0d2bb1b2d49c3e1076c15391e8c0101a3312cdbc7cdb03fdaf42f4d48"} Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.182195 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" event={"ID":"1f219932-2fca-4c2f-b44c-0d3804c427d7","Type":"ContainerStarted","Data":"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac"} Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.182239 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" event={"ID":"1f219932-2fca-4c2f-b44c-0d3804c427d7","Type":"ContainerStarted","Data":"cba19037b8b191477bfbbd427cdfad9fcf40216dfa9001eb08a428b29e86045a"} Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.182422 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.186119 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.197657 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.198474 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" podStartSLOduration=4.198459809 podStartE2EDuration="4.198459809s" podCreationTimestamp="2026-01-29 12:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:10:49.195799314 +0000 UTC m=+333.210929430" watchObservedRunningTime="2026-01-29 12:10:49.198459809 +0000 UTC m=+333.213589935" Jan 29 12:10:49 crc kubenswrapper[4993]: I0129 12:10:49.232622 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" podStartSLOduration=4.232602019 podStartE2EDuration="4.232602019s" podCreationTimestamp="2026-01-29 12:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:10:49.231621301 +0000 UTC m=+333.246751427" watchObservedRunningTime="2026-01-29 12:10:49.232602019 +0000 UTC m=+333.247732145" Jan 29 12:11:05 crc kubenswrapper[4993]: I0129 12:11:05.790281 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:11:05 crc kubenswrapper[4993]: I0129 12:11:05.792284 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" podUID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" containerName="controller-manager" containerID="cri-o://d65b919e1b133b197030e38e5835e68e7d95621cc8c162a9b9259e09b0da5f53" gracePeriod=30 Jan 29 12:11:05 crc kubenswrapper[4993]: I0129 12:11:05.803498 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:11:05 crc kubenswrapper[4993]: I0129 12:11:05.804015 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" podUID="1f219932-2fca-4c2f-b44c-0d3804c427d7" containerName="route-controller-manager" containerID="cri-o://6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac" gracePeriod=30 Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.231639 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.274468 4993 generic.go:334] "Generic (PLEG): container finished" podID="1f219932-2fca-4c2f-b44c-0d3804c427d7" containerID="6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac" exitCode=0 Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.274533 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.274551 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" event={"ID":"1f219932-2fca-4c2f-b44c-0d3804c427d7","Type":"ContainerDied","Data":"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac"} Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.274585 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb" event={"ID":"1f219932-2fca-4c2f-b44c-0d3804c427d7","Type":"ContainerDied","Data":"cba19037b8b191477bfbbd427cdfad9fcf40216dfa9001eb08a428b29e86045a"} Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.274605 4993 scope.go:117] "RemoveContainer" containerID="6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.282212 4993 generic.go:334] "Generic (PLEG): container finished" podID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" containerID="d65b919e1b133b197030e38e5835e68e7d95621cc8c162a9b9259e09b0da5f53" exitCode=0 Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.282245 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" event={"ID":"4d046304-d1f8-41e2-a8c0-885fe1aeefbd","Type":"ContainerDied","Data":"d65b919e1b133b197030e38e5835e68e7d95621cc8c162a9b9259e09b0da5f53"} Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.306715 4993 scope.go:117] "RemoveContainer" containerID="6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac" Jan 29 12:11:06 crc kubenswrapper[4993]: E0129 12:11:06.307404 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac\": container with ID starting with 6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac not found: ID does not exist" containerID="6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.307436 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac"} err="failed to get container status \"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac\": rpc error: code = NotFound desc = could not find container \"6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac\": container with ID starting with 6624dd8ce245048a59d22fdaae6391d93503df4954e0534e8cb7ecf7249e97ac not found: ID does not exist" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.358731 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert\") pod \"1f219932-2fca-4c2f-b44c-0d3804c427d7\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.358784 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config\") pod \"1f219932-2fca-4c2f-b44c-0d3804c427d7\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.358854 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca\") pod \"1f219932-2fca-4c2f-b44c-0d3804c427d7\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.358914 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvwgb\" (UniqueName: \"kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb\") pod \"1f219932-2fca-4c2f-b44c-0d3804c427d7\" (UID: \"1f219932-2fca-4c2f-b44c-0d3804c427d7\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.359931 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca" (OuterVolumeSpecName: "client-ca") pod "1f219932-2fca-4c2f-b44c-0d3804c427d7" (UID: "1f219932-2fca-4c2f-b44c-0d3804c427d7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.360008 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config" (OuterVolumeSpecName: "config") pod "1f219932-2fca-4c2f-b44c-0d3804c427d7" (UID: "1f219932-2fca-4c2f-b44c-0d3804c427d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.363865 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1f219932-2fca-4c2f-b44c-0d3804c427d7" (UID: "1f219932-2fca-4c2f-b44c-0d3804c427d7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.364108 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb" (OuterVolumeSpecName: "kube-api-access-zvwgb") pod "1f219932-2fca-4c2f-b44c-0d3804c427d7" (UID: "1f219932-2fca-4c2f-b44c-0d3804c427d7"). InnerVolumeSpecName "kube-api-access-zvwgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.426117 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.460683 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.460721 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvwgb\" (UniqueName: \"kubernetes.io/projected/1f219932-2fca-4c2f-b44c-0d3804c427d7-kube-api-access-zvwgb\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.460735 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f219932-2fca-4c2f-b44c-0d3804c427d7-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.460746 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f219932-2fca-4c2f-b44c-0d3804c427d7-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.561545 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles\") pod \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.561599 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config\") pod \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.561643 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert\") pod \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.561676 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rp4t\" (UniqueName: \"kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t\") pod \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.561703 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca\") pod \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\" (UID: \"4d046304-d1f8-41e2-a8c0-885fe1aeefbd\") " Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.562700 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca" (OuterVolumeSpecName: "client-ca") pod "4d046304-d1f8-41e2-a8c0-885fe1aeefbd" (UID: "4d046304-d1f8-41e2-a8c0-885fe1aeefbd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.562783 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4d046304-d1f8-41e2-a8c0-885fe1aeefbd" (UID: "4d046304-d1f8-41e2-a8c0-885fe1aeefbd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.562816 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config" (OuterVolumeSpecName: "config") pod "4d046304-d1f8-41e2-a8c0-885fe1aeefbd" (UID: "4d046304-d1f8-41e2-a8c0-885fe1aeefbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.564699 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t" (OuterVolumeSpecName: "kube-api-access-5rp4t") pod "4d046304-d1f8-41e2-a8c0-885fe1aeefbd" (UID: "4d046304-d1f8-41e2-a8c0-885fe1aeefbd"). InnerVolumeSpecName "kube-api-access-5rp4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.565110 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4d046304-d1f8-41e2-a8c0-885fe1aeefbd" (UID: "4d046304-d1f8-41e2-a8c0-885fe1aeefbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.604854 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.609218 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-l8mmb"] Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.663637 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.663683 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.663697 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rp4t\" (UniqueName: \"kubernetes.io/projected/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-kube-api-access-5rp4t\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.663711 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:06 crc kubenswrapper[4993]: I0129 12:11:06.663723 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d046304-d1f8-41e2-a8c0-885fe1aeefbd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.195583 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f219932-2fca-4c2f-b44c-0d3804c427d7" path="/var/lib/kubelet/pods/1f219932-2fca-4c2f-b44c-0d3804c427d7/volumes" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.288416 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" event={"ID":"4d046304-d1f8-41e2-a8c0-885fe1aeefbd","Type":"ContainerDied","Data":"6f5d9fa0d2bb1b2d49c3e1076c15391e8c0101a3312cdbc7cdb03fdaf42f4d48"} Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.288435 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75795fb4fc-c2vxg" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.288566 4993 scope.go:117] "RemoveContainer" containerID="d65b919e1b133b197030e38e5835e68e7d95621cc8c162a9b9259e09b0da5f53" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.314315 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.325660 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-75795fb4fc-c2vxg"] Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760246 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:07 crc kubenswrapper[4993]: E0129 12:11:07.760475 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f219932-2fca-4c2f-b44c-0d3804c427d7" containerName="route-controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760488 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f219932-2fca-4c2f-b44c-0d3804c427d7" containerName="route-controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: E0129 12:11:07.760509 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" containerName="controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760516 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" containerName="controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760614 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" containerName="controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760623 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f219932-2fca-4c2f-b44c-0d3804c427d7" containerName="route-controller-manager" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.760981 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.764004 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.764675 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.766095 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.766385 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.766644 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.766913 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.767071 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.767175 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.767289 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.767391 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.768444 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.768840 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.768902 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.772998 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.776233 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.777197 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.779956 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877146 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pkql\" (UniqueName: \"kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877216 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877271 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877297 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877362 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877395 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877422 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877533 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.877573 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh2jk\" (UniqueName: \"kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978521 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978603 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978643 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978688 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978733 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978802 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978836 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh2jk\" (UniqueName: \"kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978875 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.978904 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pkql\" (UniqueName: \"kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.980578 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.980647 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.980784 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.981096 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.981356 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.983137 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.983136 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:07 crc kubenswrapper[4993]: I0129 12:11:07.998896 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pkql\" (UniqueName: \"kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql\") pod \"controller-manager-6dd67f4bb8-tb8w2\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:08 crc kubenswrapper[4993]: I0129 12:11:08.001469 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh2jk\" (UniqueName: \"kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk\") pod \"route-controller-manager-77467c85cc-nsv9j\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:08 crc kubenswrapper[4993]: I0129 12:11:08.081044 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:08 crc kubenswrapper[4993]: I0129 12:11:08.093104 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:08 crc kubenswrapper[4993]: I0129 12:11:08.326360 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:08 crc kubenswrapper[4993]: W0129 12:11:08.330453 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcea4402c_9f99_44ec_b86f_e658c08738de.slice/crio-f9eb37805a9848144bf5aa0a60e51fcca6b6b9371d93edb1226558cdc475d654 WatchSource:0}: Error finding container f9eb37805a9848144bf5aa0a60e51fcca6b6b9371d93edb1226558cdc475d654: Status 404 returned error can't find the container with id f9eb37805a9848144bf5aa0a60e51fcca6b6b9371d93edb1226558cdc475d654 Jan 29 12:11:08 crc kubenswrapper[4993]: I0129 12:11:08.488381 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:08 crc kubenswrapper[4993]: W0129 12:11:08.493437 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod104a86fa_3e76_4438_b6de_d9f067ed85fd.slice/crio-4fc323287d129c38cd2f3d6faadc30d6a2a5edabb387983d8463a25dcbcb5bfe WatchSource:0}: Error finding container 4fc323287d129c38cd2f3d6faadc30d6a2a5edabb387983d8463a25dcbcb5bfe: Status 404 returned error can't find the container with id 4fc323287d129c38cd2f3d6faadc30d6a2a5edabb387983d8463a25dcbcb5bfe Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.197791 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d046304-d1f8-41e2-a8c0-885fe1aeefbd" path="/var/lib/kubelet/pods/4d046304-d1f8-41e2-a8c0-885fe1aeefbd/volumes" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.304693 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" event={"ID":"104a86fa-3e76-4438-b6de-d9f067ed85fd","Type":"ContainerStarted","Data":"11a72dbd44dbd01cae984d0463cbc8a22d2fca4a885916cb04ec627789a213ed"} Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.305294 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" event={"ID":"104a86fa-3e76-4438-b6de-d9f067ed85fd","Type":"ContainerStarted","Data":"4fc323287d129c38cd2f3d6faadc30d6a2a5edabb387983d8463a25dcbcb5bfe"} Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.305327 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.306599 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" event={"ID":"cea4402c-9f99-44ec-b86f-e658c08738de","Type":"ContainerStarted","Data":"30fb6afe7c9b12c73e217e2463636dd8b93005e1816e46e7cb64efcf5e431d3d"} Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.306665 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" event={"ID":"cea4402c-9f99-44ec-b86f-e658c08738de","Type":"ContainerStarted","Data":"f9eb37805a9848144bf5aa0a60e51fcca6b6b9371d93edb1226558cdc475d654"} Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.306791 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.309892 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.311981 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.346018 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" podStartSLOduration=4.34600219 podStartE2EDuration="4.34600219s" podCreationTimestamp="2026-01-29 12:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:09.323635165 +0000 UTC m=+353.338765291" watchObservedRunningTime="2026-01-29 12:11:09.34600219 +0000 UTC m=+353.361132316" Jan 29 12:11:09 crc kubenswrapper[4993]: I0129 12:11:09.368560 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" podStartSLOduration=4.36854105 podStartE2EDuration="4.36854105s" podCreationTimestamp="2026-01-29 12:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:09.367053637 +0000 UTC m=+353.382183763" watchObservedRunningTime="2026-01-29 12:11:09.36854105 +0000 UTC m=+353.383671176" Jan 29 12:11:18 crc kubenswrapper[4993]: I0129 12:11:18.156218 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:11:18 crc kubenswrapper[4993]: I0129 12:11:18.156883 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:11:25 crc kubenswrapper[4993]: I0129 12:11:25.796366 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:25 crc kubenswrapper[4993]: I0129 12:11:25.797066 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" podUID="104a86fa-3e76-4438-b6de-d9f067ed85fd" containerName="controller-manager" containerID="cri-o://11a72dbd44dbd01cae984d0463cbc8a22d2fca4a885916cb04ec627789a213ed" gracePeriod=30 Jan 29 12:11:25 crc kubenswrapper[4993]: I0129 12:11:25.917912 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:25 crc kubenswrapper[4993]: I0129 12:11:25.918455 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" podUID="cea4402c-9f99-44ec-b86f-e658c08738de" containerName="route-controller-manager" containerID="cri-o://30fb6afe7c9b12c73e217e2463636dd8b93005e1816e46e7cb64efcf5e431d3d" gracePeriod=30 Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.397252 4993 generic.go:334] "Generic (PLEG): container finished" podID="104a86fa-3e76-4438-b6de-d9f067ed85fd" containerID="11a72dbd44dbd01cae984d0463cbc8a22d2fca4a885916cb04ec627789a213ed" exitCode=0 Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.397538 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" event={"ID":"104a86fa-3e76-4438-b6de-d9f067ed85fd","Type":"ContainerDied","Data":"11a72dbd44dbd01cae984d0463cbc8a22d2fca4a885916cb04ec627789a213ed"} Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.398752 4993 generic.go:334] "Generic (PLEG): container finished" podID="cea4402c-9f99-44ec-b86f-e658c08738de" containerID="30fb6afe7c9b12c73e217e2463636dd8b93005e1816e46e7cb64efcf5e431d3d" exitCode=0 Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.398780 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" event={"ID":"cea4402c-9f99-44ec-b86f-e658c08738de","Type":"ContainerDied","Data":"30fb6afe7c9b12c73e217e2463636dd8b93005e1816e46e7cb64efcf5e431d3d"} Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.871640 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:26 crc kubenswrapper[4993]: I0129 12:11:26.907654 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013754 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca\") pod \"cea4402c-9f99-44ec-b86f-e658c08738de\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013822 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert\") pod \"104a86fa-3e76-4438-b6de-d9f067ed85fd\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013847 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca\") pod \"104a86fa-3e76-4438-b6de-d9f067ed85fd\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013904 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles\") pod \"104a86fa-3e76-4438-b6de-d9f067ed85fd\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013936 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh2jk\" (UniqueName: \"kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk\") pod \"cea4402c-9f99-44ec-b86f-e658c08738de\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.013991 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config\") pod \"104a86fa-3e76-4438-b6de-d9f067ed85fd\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.014078 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pkql\" (UniqueName: \"kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql\") pod \"104a86fa-3e76-4438-b6de-d9f067ed85fd\" (UID: \"104a86fa-3e76-4438-b6de-d9f067ed85fd\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.014867 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca" (OuterVolumeSpecName: "client-ca") pod "cea4402c-9f99-44ec-b86f-e658c08738de" (UID: "cea4402c-9f99-44ec-b86f-e658c08738de"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015080 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "104a86fa-3e76-4438-b6de-d9f067ed85fd" (UID: "104a86fa-3e76-4438-b6de-d9f067ed85fd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015322 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca" (OuterVolumeSpecName: "client-ca") pod "104a86fa-3e76-4438-b6de-d9f067ed85fd" (UID: "104a86fa-3e76-4438-b6de-d9f067ed85fd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015587 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config\") pod \"cea4402c-9f99-44ec-b86f-e658c08738de\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015640 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert\") pod \"cea4402c-9f99-44ec-b86f-e658c08738de\" (UID: \"cea4402c-9f99-44ec-b86f-e658c08738de\") " Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015882 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config" (OuterVolumeSpecName: "config") pod "104a86fa-3e76-4438-b6de-d9f067ed85fd" (UID: "104a86fa-3e76-4438-b6de-d9f067ed85fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015911 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015928 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.015941 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.019300 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config" (OuterVolumeSpecName: "config") pod "cea4402c-9f99-44ec-b86f-e658c08738de" (UID: "cea4402c-9f99-44ec-b86f-e658c08738de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.022032 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql" (OuterVolumeSpecName: "kube-api-access-7pkql") pod "104a86fa-3e76-4438-b6de-d9f067ed85fd" (UID: "104a86fa-3e76-4438-b6de-d9f067ed85fd"). InnerVolumeSpecName "kube-api-access-7pkql". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.022050 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cea4402c-9f99-44ec-b86f-e658c08738de" (UID: "cea4402c-9f99-44ec-b86f-e658c08738de"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.022833 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk" (OuterVolumeSpecName: "kube-api-access-fh2jk") pod "cea4402c-9f99-44ec-b86f-e658c08738de" (UID: "cea4402c-9f99-44ec-b86f-e658c08738de"). InnerVolumeSpecName "kube-api-access-fh2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.029704 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "104a86fa-3e76-4438-b6de-d9f067ed85fd" (UID: "104a86fa-3e76-4438-b6de-d9f067ed85fd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117442 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh2jk\" (UniqueName: \"kubernetes.io/projected/cea4402c-9f99-44ec-b86f-e658c08738de-kube-api-access-fh2jk\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117491 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/104a86fa-3e76-4438-b6de-d9f067ed85fd-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117501 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pkql\" (UniqueName: \"kubernetes.io/projected/104a86fa-3e76-4438-b6de-d9f067ed85fd-kube-api-access-7pkql\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117510 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea4402c-9f99-44ec-b86f-e658c08738de-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117519 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea4402c-9f99-44ec-b86f-e658c08738de-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.117528 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/104a86fa-3e76-4438-b6de-d9f067ed85fd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.404309 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.404330 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2" event={"ID":"104a86fa-3e76-4438-b6de-d9f067ed85fd","Type":"ContainerDied","Data":"4fc323287d129c38cd2f3d6faadc30d6a2a5edabb387983d8463a25dcbcb5bfe"} Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.404381 4993 scope.go:117] "RemoveContainer" containerID="11a72dbd44dbd01cae984d0463cbc8a22d2fca4a885916cb04ec627789a213ed" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.406014 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" event={"ID":"cea4402c-9f99-44ec-b86f-e658c08738de","Type":"ContainerDied","Data":"f9eb37805a9848144bf5aa0a60e51fcca6b6b9371d93edb1226558cdc475d654"} Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.406321 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.423225 4993 scope.go:117] "RemoveContainer" containerID="30fb6afe7c9b12c73e217e2463636dd8b93005e1816e46e7cb64efcf5e431d3d" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.436634 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.444284 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6dd67f4bb8-tb8w2"] Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.447567 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.450830 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77467c85cc-nsv9j"] Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.772641 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:27 crc kubenswrapper[4993]: E0129 12:11:27.772885 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea4402c-9f99-44ec-b86f-e658c08738de" containerName="route-controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.772904 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea4402c-9f99-44ec-b86f-e658c08738de" containerName="route-controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: E0129 12:11:27.772919 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104a86fa-3e76-4438-b6de-d9f067ed85fd" containerName="controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.772926 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="104a86fa-3e76-4438-b6de-d9f067ed85fd" containerName="controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.773055 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea4402c-9f99-44ec-b86f-e658c08738de" containerName="route-controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.773069 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="104a86fa-3e76-4438-b6de-d9f067ed85fd" containerName="controller-manager" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.773527 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.777794 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.777881 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.777985 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.777994 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.778134 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.778205 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.790866 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.927891 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.927961 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.928040 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgtt9\" (UniqueName: \"kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:27 crc kubenswrapper[4993]: I0129 12:11:27.928117 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.029376 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgtt9\" (UniqueName: \"kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.029486 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.029645 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.031454 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.031500 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.032267 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.040291 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.048952 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgtt9\" (UniqueName: \"kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9\") pod \"route-controller-manager-646585b7f6-kct7h\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.090885 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.290207 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:28 crc kubenswrapper[4993]: I0129 12:11:28.414950 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" event={"ID":"31eb40cd-db58-4a9a-9b30-feafb83122ec","Type":"ContainerStarted","Data":"2f2c299c17a41550a295d6b00d537425f69f6f9f59594531992338036c288bdc"} Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.199143 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="104a86fa-3e76-4438-b6de-d9f067ed85fd" path="/var/lib/kubelet/pods/104a86fa-3e76-4438-b6de-d9f067ed85fd/volumes" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.200674 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea4402c-9f99-44ec-b86f-e658c08738de" path="/var/lib/kubelet/pods/cea4402c-9f99-44ec-b86f-e658c08738de/volumes" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.423558 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" event={"ID":"31eb40cd-db58-4a9a-9b30-feafb83122ec","Type":"ContainerStarted","Data":"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822"} Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.424425 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.430325 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.444249 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" podStartSLOduration=4.444231271 podStartE2EDuration="4.444231271s" podCreationTimestamp="2026-01-29 12:11:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:29.441923266 +0000 UTC m=+373.457053422" watchObservedRunningTime="2026-01-29 12:11:29.444231271 +0000 UTC m=+373.459361417" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.774914 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.775703 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.777340 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.777346 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.778319 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.778403 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.778582 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.778770 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.785143 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.788670 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.860077 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.860215 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.860255 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhqzt\" (UniqueName: \"kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.860324 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.860411 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.961875 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.961962 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.961988 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhqzt\" (UniqueName: \"kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.962028 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.962074 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.963291 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.963784 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.964213 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.968464 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:29 crc kubenswrapper[4993]: I0129 12:11:29.978652 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhqzt\" (UniqueName: \"kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt\") pod \"controller-manager-59fb48b9cf-brw5b\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:30 crc kubenswrapper[4993]: I0129 12:11:30.094420 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:30 crc kubenswrapper[4993]: I0129 12:11:30.474780 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:30 crc kubenswrapper[4993]: W0129 12:11:30.482358 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1dcf88e6_e412_4ee0_b76e_841b5c9706a6.slice/crio-fdda7bb2d39ed6eff3f94a6ed85b3ae1c9dfab4fad803d40a7a1178cdef763dd WatchSource:0}: Error finding container fdda7bb2d39ed6eff3f94a6ed85b3ae1c9dfab4fad803d40a7a1178cdef763dd: Status 404 returned error can't find the container with id fdda7bb2d39ed6eff3f94a6ed85b3ae1c9dfab4fad803d40a7a1178cdef763dd Jan 29 12:11:31 crc kubenswrapper[4993]: I0129 12:11:31.434138 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" event={"ID":"1dcf88e6-e412-4ee0-b76e-841b5c9706a6","Type":"ContainerStarted","Data":"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1"} Jan 29 12:11:31 crc kubenswrapper[4993]: I0129 12:11:31.434224 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" event={"ID":"1dcf88e6-e412-4ee0-b76e-841b5c9706a6","Type":"ContainerStarted","Data":"fdda7bb2d39ed6eff3f94a6ed85b3ae1c9dfab4fad803d40a7a1178cdef763dd"} Jan 29 12:11:31 crc kubenswrapper[4993]: I0129 12:11:31.454831 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" podStartSLOduration=6.454815793 podStartE2EDuration="6.454815793s" podCreationTimestamp="2026-01-29 12:11:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:31.451391756 +0000 UTC m=+375.466521882" watchObservedRunningTime="2026-01-29 12:11:31.454815793 +0000 UTC m=+375.469945919" Jan 29 12:11:32 crc kubenswrapper[4993]: I0129 12:11:32.438703 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:32 crc kubenswrapper[4993]: I0129 12:11:32.442790 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:41 crc kubenswrapper[4993]: I0129 12:11:41.071776 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:11:45 crc kubenswrapper[4993]: I0129 12:11:45.820917 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:45 crc kubenswrapper[4993]: I0129 12:11:45.822402 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" podUID="31eb40cd-db58-4a9a-9b30-feafb83122ec" containerName="route-controller-manager" containerID="cri-o://dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822" gracePeriod=30 Jan 29 12:11:45 crc kubenswrapper[4993]: I0129 12:11:45.824852 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:45 crc kubenswrapper[4993]: I0129 12:11:45.825103 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" podUID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" containerName="controller-manager" containerID="cri-o://5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1" gracePeriod=30 Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.352798 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.466239 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgtt9\" (UniqueName: \"kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9\") pod \"31eb40cd-db58-4a9a-9b30-feafb83122ec\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.466311 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca\") pod \"31eb40cd-db58-4a9a-9b30-feafb83122ec\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.466442 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert\") pod \"31eb40cd-db58-4a9a-9b30-feafb83122ec\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.466499 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config\") pod \"31eb40cd-db58-4a9a-9b30-feafb83122ec\" (UID: \"31eb40cd-db58-4a9a-9b30-feafb83122ec\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.467057 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca" (OuterVolumeSpecName: "client-ca") pod "31eb40cd-db58-4a9a-9b30-feafb83122ec" (UID: "31eb40cd-db58-4a9a-9b30-feafb83122ec"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.467433 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config" (OuterVolumeSpecName: "config") pod "31eb40cd-db58-4a9a-9b30-feafb83122ec" (UID: "31eb40cd-db58-4a9a-9b30-feafb83122ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.472413 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9" (OuterVolumeSpecName: "kube-api-access-mgtt9") pod "31eb40cd-db58-4a9a-9b30-feafb83122ec" (UID: "31eb40cd-db58-4a9a-9b30-feafb83122ec"). InnerVolumeSpecName "kube-api-access-mgtt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.472614 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "31eb40cd-db58-4a9a-9b30-feafb83122ec" (UID: "31eb40cd-db58-4a9a-9b30-feafb83122ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.496310 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.509086 4993 generic.go:334] "Generic (PLEG): container finished" podID="31eb40cd-db58-4a9a-9b30-feafb83122ec" containerID="dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822" exitCode=0 Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.509176 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" event={"ID":"31eb40cd-db58-4a9a-9b30-feafb83122ec","Type":"ContainerDied","Data":"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822"} Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.509262 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" event={"ID":"31eb40cd-db58-4a9a-9b30-feafb83122ec","Type":"ContainerDied","Data":"2f2c299c17a41550a295d6b00d537425f69f6f9f59594531992338036c288bdc"} Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.509287 4993 scope.go:117] "RemoveContainer" containerID="dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.509540 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.514639 4993 generic.go:334] "Generic (PLEG): container finished" podID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" containerID="5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1" exitCode=0 Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.514880 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.514901 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" event={"ID":"1dcf88e6-e412-4ee0-b76e-841b5c9706a6","Type":"ContainerDied","Data":"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1"} Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.515953 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59fb48b9cf-brw5b" event={"ID":"1dcf88e6-e412-4ee0-b76e-841b5c9706a6","Type":"ContainerDied","Data":"fdda7bb2d39ed6eff3f94a6ed85b3ae1c9dfab4fad803d40a7a1178cdef763dd"} Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.527211 4993 scope.go:117] "RemoveContainer" containerID="dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822" Jan 29 12:11:46 crc kubenswrapper[4993]: E0129 12:11:46.530554 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822\": container with ID starting with dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822 not found: ID does not exist" containerID="dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.530604 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822"} err="failed to get container status \"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822\": rpc error: code = NotFound desc = could not find container \"dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822\": container with ID starting with dcf318927db67f2ccb8d092ce83996d7451927fdde69d76cd01f1e73c293d822 not found: ID does not exist" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.530639 4993 scope.go:117] "RemoveContainer" containerID="5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.546756 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.552999 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-646585b7f6-kct7h"] Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.553477 4993 scope.go:117] "RemoveContainer" containerID="5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1" Jan 29 12:11:46 crc kubenswrapper[4993]: E0129 12:11:46.554026 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1\": container with ID starting with 5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1 not found: ID does not exist" containerID="5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.554124 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1"} err="failed to get container status \"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1\": rpc error: code = NotFound desc = could not find container \"5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1\": container with ID starting with 5d83d529a2aca88d46c3331507b1fab31ba84a881781f993a9bc990b0347a9e1 not found: ID does not exist" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.567784 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca\") pod \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568033 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config\") pod \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568218 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert\") pod \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568378 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhqzt\" (UniqueName: \"kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt\") pod \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568495 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles\") pod \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\" (UID: \"1dcf88e6-e412-4ee0-b76e-841b5c9706a6\") " Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568801 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31eb40cd-db58-4a9a-9b30-feafb83122ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568902 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.568986 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgtt9\" (UniqueName: \"kubernetes.io/projected/31eb40cd-db58-4a9a-9b30-feafb83122ec-kube-api-access-mgtt9\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.569089 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31eb40cd-db58-4a9a-9b30-feafb83122ec-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.569301 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "1dcf88e6-e412-4ee0-b76e-841b5c9706a6" (UID: "1dcf88e6-e412-4ee0-b76e-841b5c9706a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.569337 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1dcf88e6-e412-4ee0-b76e-841b5c9706a6" (UID: "1dcf88e6-e412-4ee0-b76e-841b5c9706a6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.569320 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config" (OuterVolumeSpecName: "config") pod "1dcf88e6-e412-4ee0-b76e-841b5c9706a6" (UID: "1dcf88e6-e412-4ee0-b76e-841b5c9706a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.571740 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt" (OuterVolumeSpecName: "kube-api-access-fhqzt") pod "1dcf88e6-e412-4ee0-b76e-841b5c9706a6" (UID: "1dcf88e6-e412-4ee0-b76e-841b5c9706a6"). InnerVolumeSpecName "kube-api-access-fhqzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.571724 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1dcf88e6-e412-4ee0-b76e-841b5c9706a6" (UID: "1dcf88e6-e412-4ee0-b76e-841b5c9706a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.670098 4993 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.670141 4993 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.670150 4993 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.670161 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhqzt\" (UniqueName: \"kubernetes.io/projected/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-kube-api-access-fhqzt\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.670173 4993 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dcf88e6-e412-4ee0-b76e-841b5c9706a6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.845885 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:46 crc kubenswrapper[4993]: I0129 12:11:46.849228 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-59fb48b9cf-brw5b"] Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.197643 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" path="/var/lib/kubelet/pods/1dcf88e6-e412-4ee0-b76e-841b5c9706a6/volumes" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.198603 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31eb40cd-db58-4a9a-9b30-feafb83122ec" path="/var/lib/kubelet/pods/31eb40cd-db58-4a9a-9b30-feafb83122ec/volumes" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.786484 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb"] Jan 29 12:11:47 crc kubenswrapper[4993]: E0129 12:11:47.786848 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31eb40cd-db58-4a9a-9b30-feafb83122ec" containerName="route-controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.786871 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="31eb40cd-db58-4a9a-9b30-feafb83122ec" containerName="route-controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: E0129 12:11:47.786884 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" containerName="controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.786893 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" containerName="controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.787008 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="31eb40cd-db58-4a9a-9b30-feafb83122ec" containerName="route-controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.787031 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dcf88e6-e412-4ee0-b76e-841b5c9706a6" containerName="controller-manager" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.787650 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.791395 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5bcfb4976d-shqbn"] Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.791661 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.792373 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.792833 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.792865 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.794285 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.794476 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.794591 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.794718 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.794982 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.795083 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.800293 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.800294 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.801842 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.802673 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bcfb4976d-shqbn"] Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.803738 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.807527 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb"] Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884687 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-proxy-ca-bundles\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884736 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-config\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884766 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcwmj\" (UniqueName: \"kubernetes.io/projected/490711a8-3fa9-47db-b78d-3025ab91f88a-kube-api-access-lcwmj\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884808 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnxbc\" (UniqueName: \"kubernetes.io/projected/c445d83b-6867-4bfd-bda1-0d20f707e2dc-kube-api-access-nnxbc\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884830 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-client-ca\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884847 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490711a8-3fa9-47db-b78d-3025ab91f88a-serving-cert\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884866 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c445d83b-6867-4bfd-bda1-0d20f707e2dc-serving-cert\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884897 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-client-ca\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.884925 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-config\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.985916 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-config\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.985963 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-proxy-ca-bundles\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.985989 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-config\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986009 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcwmj\" (UniqueName: \"kubernetes.io/projected/490711a8-3fa9-47db-b78d-3025ab91f88a-kube-api-access-lcwmj\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986051 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnxbc\" (UniqueName: \"kubernetes.io/projected/c445d83b-6867-4bfd-bda1-0d20f707e2dc-kube-api-access-nnxbc\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986073 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-client-ca\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986090 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490711a8-3fa9-47db-b78d-3025ab91f88a-serving-cert\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986108 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c445d83b-6867-4bfd-bda1-0d20f707e2dc-serving-cert\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.986137 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-client-ca\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.987060 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-client-ca\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.987288 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-config\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.987592 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-config\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.987635 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c445d83b-6867-4bfd-bda1-0d20f707e2dc-proxy-ca-bundles\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.987781 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490711a8-3fa9-47db-b78d-3025ab91f88a-client-ca\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.990508 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c445d83b-6867-4bfd-bda1-0d20f707e2dc-serving-cert\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:47 crc kubenswrapper[4993]: I0129 12:11:47.991084 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490711a8-3fa9-47db-b78d-3025ab91f88a-serving-cert\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.004958 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnxbc\" (UniqueName: \"kubernetes.io/projected/c445d83b-6867-4bfd-bda1-0d20f707e2dc-kube-api-access-nnxbc\") pod \"controller-manager-5bcfb4976d-shqbn\" (UID: \"c445d83b-6867-4bfd-bda1-0d20f707e2dc\") " pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.005684 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcwmj\" (UniqueName: \"kubernetes.io/projected/490711a8-3fa9-47db-b78d-3025ab91f88a-kube-api-access-lcwmj\") pod \"route-controller-manager-5c7b75974f-f6jcb\" (UID: \"490711a8-3fa9-47db-b78d-3025ab91f88a\") " pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.120833 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.129469 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.158597 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.158670 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.550560 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb"] Jan 29 12:11:48 crc kubenswrapper[4993]: I0129 12:11:48.588691 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bcfb4976d-shqbn"] Jan 29 12:11:48 crc kubenswrapper[4993]: W0129 12:11:48.591914 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc445d83b_6867_4bfd_bda1_0d20f707e2dc.slice/crio-369def4dd120140625e786acaae6f9465c76f3f1e98a6401c64176ada9ee78fb WatchSource:0}: Error finding container 369def4dd120140625e786acaae6f9465c76f3f1e98a6401c64176ada9ee78fb: Status 404 returned error can't find the container with id 369def4dd120140625e786acaae6f9465c76f3f1e98a6401c64176ada9ee78fb Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.533561 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" event={"ID":"c445d83b-6867-4bfd-bda1-0d20f707e2dc","Type":"ContainerStarted","Data":"c2d6415e178140a7d113d1d61b43c8ffdf9d5832eb22891c4aa11a58d1dca2b0"} Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.534053 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.534075 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" event={"ID":"c445d83b-6867-4bfd-bda1-0d20f707e2dc","Type":"ContainerStarted","Data":"369def4dd120140625e786acaae6f9465c76f3f1e98a6401c64176ada9ee78fb"} Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.536432 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" event={"ID":"490711a8-3fa9-47db-b78d-3025ab91f88a","Type":"ContainerStarted","Data":"3dc6eae6219cf53d1f32e063a024cdb4d17d84c92900a51cc139d3fa80b3b53c"} Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.536472 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" event={"ID":"490711a8-3fa9-47db-b78d-3025ab91f88a","Type":"ContainerStarted","Data":"6638391ba1ed043527fb1425cff3b53a47ed9d247d341da44f0beec1384a84f0"} Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.536671 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.537895 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.542964 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" Jan 29 12:11:49 crc kubenswrapper[4993]: I0129 12:11:49.567582 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5bcfb4976d-shqbn" podStartSLOduration=4.567556949 podStartE2EDuration="4.567556949s" podCreationTimestamp="2026-01-29 12:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:49.550939883 +0000 UTC m=+393.566070019" watchObservedRunningTime="2026-01-29 12:11:49.567556949 +0000 UTC m=+393.582687085" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.099376 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" containerID="cri-o://8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c" gracePeriod=15 Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.596891 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.616498 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c7b75974f-f6jcb" podStartSLOduration=21.616479489 podStartE2EDuration="21.616479489s" podCreationTimestamp="2026-01-29 12:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:11:49.591593232 +0000 UTC m=+393.606723358" watchObservedRunningTime="2026-01-29 12:12:06.616479489 +0000 UTC m=+410.631609605" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.626721 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5c7b884f55-pbj28"] Jan 29 12:12:06 crc kubenswrapper[4993]: E0129 12:12:06.626994 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.627010 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.627135 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerName="oauth-openshift" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.627633 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.646942 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5c7b884f55-pbj28"] Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.674471 4993 generic.go:334] "Generic (PLEG): container finished" podID="887d412a-3e41-473f-9954-4fcfccf6b2ea" containerID="8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c" exitCode=0 Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.674777 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" event={"ID":"887d412a-3e41-473f-9954-4fcfccf6b2ea","Type":"ContainerDied","Data":"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c"} Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.674999 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" event={"ID":"887d412a-3e41-473f-9954-4fcfccf6b2ea","Type":"ContainerDied","Data":"1e6729faa1ae6b24e4f34106769ab1384e011f7e7d99fd214720307aeff6e84c"} Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.674820 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9zsmx" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.675122 4993 scope.go:117] "RemoveContainer" containerID="8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.697521 4993 scope.go:117] "RemoveContainer" containerID="8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c" Jan 29 12:12:06 crc kubenswrapper[4993]: E0129 12:12:06.699520 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c\": container with ID starting with 8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c not found: ID does not exist" containerID="8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.699586 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c"} err="failed to get container status \"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c\": rpc error: code = NotFound desc = could not find container \"8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c\": container with ID starting with 8d17a4eb8724427d3ac0cb73473245b4b4a7b11fe3f8bd8f7fba1b0de214c64c not found: ID does not exist" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715758 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715830 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715858 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715876 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715863 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715905 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715927 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715943 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715968 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.715988 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716008 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716025 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716084 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6czdj\" (UniqueName: \"kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716108 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716132 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig\") pod \"887d412a-3e41-473f-9954-4fcfccf6b2ea\" (UID: \"887d412a-3e41-473f-9954-4fcfccf6b2ea\") " Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716321 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716362 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716387 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716414 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-policies\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716441 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-session\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716463 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-login\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716485 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-dir\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716506 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716527 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p97zz\" (UniqueName: \"kubernetes.io/projected/3b6e89df-6af8-4fce-b04f-e2a286f05550-kube-api-access-p97zz\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716554 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-error\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716581 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716626 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716650 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716672 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.716755 4993 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.717596 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.718007 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.718041 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.718492 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.723914 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.723926 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj" (OuterVolumeSpecName: "kube-api-access-6czdj") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "kube-api-access-6czdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.725772 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.726521 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.726780 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.729876 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.731524 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.731775 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.731929 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "887d412a-3e41-473f-9954-4fcfccf6b2ea" (UID: "887d412a-3e41-473f-9954-4fcfccf6b2ea"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.817901 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-error\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.817945 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.817978 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.817997 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818015 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818050 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818074 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818091 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818112 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-policies\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818132 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-session\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818149 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-login\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818169 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-dir\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818218 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818242 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p97zz\" (UniqueName: \"kubernetes.io/projected/3b6e89df-6af8-4fce-b04f-e2a286f05550-kube-api-access-p97zz\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818292 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818308 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6czdj\" (UniqueName: \"kubernetes.io/projected/887d412a-3e41-473f-9954-4fcfccf6b2ea-kube-api-access-6czdj\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818320 4993 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818332 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818343 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818355 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818367 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818378 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818851 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818865 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818879 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818894 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.818906 4993 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d412a-3e41-473f-9954-4fcfccf6b2ea-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.819457 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.819881 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-policies\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.820706 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.820853 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3b6e89df-6af8-4fce-b04f-e2a286f05550-audit-dir\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.820946 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.821215 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-error\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.821669 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.822163 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.822278 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.823113 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.824167 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-login\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.825019 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-system-session\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.825665 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3b6e89df-6af8-4fce-b04f-e2a286f05550-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.836827 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p97zz\" (UniqueName: \"kubernetes.io/projected/3b6e89df-6af8-4fce-b04f-e2a286f05550-kube-api-access-p97zz\") pod \"oauth-openshift-5c7b884f55-pbj28\" (UID: \"3b6e89df-6af8-4fce-b04f-e2a286f05550\") " pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:06 crc kubenswrapper[4993]: I0129 12:12:06.985288 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.006017 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.010908 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9zsmx"] Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.203347 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887d412a-3e41-473f-9954-4fcfccf6b2ea" path="/var/lib/kubelet/pods/887d412a-3e41-473f-9954-4fcfccf6b2ea/volumes" Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.392937 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5c7b884f55-pbj28"] Jan 29 12:12:07 crc kubenswrapper[4993]: W0129 12:12:07.399259 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6e89df_6af8_4fce_b04f_e2a286f05550.slice/crio-69bed064dc47651d6d3f189e68ffa18c9b0d6f91b2729290990caebb118cc66a WatchSource:0}: Error finding container 69bed064dc47651d6d3f189e68ffa18c9b0d6f91b2729290990caebb118cc66a: Status 404 returned error can't find the container with id 69bed064dc47651d6d3f189e68ffa18c9b0d6f91b2729290990caebb118cc66a Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.686172 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" event={"ID":"3b6e89df-6af8-4fce-b04f-e2a286f05550","Type":"ContainerStarted","Data":"98f2a90f1394ed41d942a169fe2a16e43d6d8e10fdbeff4df06fddd1fdfee914"} Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.686255 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" event={"ID":"3b6e89df-6af8-4fce-b04f-e2a286f05550","Type":"ContainerStarted","Data":"69bed064dc47651d6d3f189e68ffa18c9b0d6f91b2729290990caebb118cc66a"} Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.686762 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.713875 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" podStartSLOduration=26.713851364 podStartE2EDuration="26.713851364s" podCreationTimestamp="2026-01-29 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:12:07.709876512 +0000 UTC m=+411.725006638" watchObservedRunningTime="2026-01-29 12:12:07.713851364 +0000 UTC m=+411.728981490" Jan 29 12:12:07 crc kubenswrapper[4993]: I0129 12:12:07.985866 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5c7b884f55-pbj28" Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.156168 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.158268 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.158479 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.159727 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.159881 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89" gracePeriod=600 Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.740575 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89" exitCode=0 Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.740636 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89"} Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.740911 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b"} Jan 29 12:12:18 crc kubenswrapper[4993]: I0129 12:12:18.740936 4993 scope.go:117] "RemoveContainer" containerID="9abf81b8ca6779831bd58aaeb82749834c0de645a81c49da2beb167cc52f6018" Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.854286 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.855340 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zzw5p" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="registry-server" containerID="cri-o://5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b" gracePeriod=30 Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.868433 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.869105 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dlzds" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="registry-server" containerID="cri-o://549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5" gracePeriod=30 Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.878309 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.881834 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" containerID="cri-o://ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f" gracePeriod=30 Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.893131 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.893496 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n7xrp" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="registry-server" containerID="cri-o://4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed" gracePeriod=30 Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.900137 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvrlt"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.901031 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.913297 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.913545 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9nh5h" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="registry-server" containerID="cri-o://78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede" gracePeriod=30 Jan 29 12:12:27 crc kubenswrapper[4993]: I0129 12:12:27.933588 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvrlt"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.036516 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.036661 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.036692 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-977cz\" (UniqueName: \"kubernetes.io/projected/2f0e3bcf-c987-4ef4-9055-31b947d2807c-kube-api-access-977cz\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.137578 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.137624 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-977cz\" (UniqueName: \"kubernetes.io/projected/2f0e3bcf-c987-4ef4-9055-31b947d2807c-kube-api-access-977cz\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.138312 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.139157 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.146576 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f0e3bcf-c987-4ef4-9055-31b947d2807c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.155553 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-977cz\" (UniqueName: \"kubernetes.io/projected/2f0e3bcf-c987-4ef4-9055-31b947d2807c-kube-api-access-977cz\") pod \"marketplace-operator-79b997595-qvrlt\" (UID: \"2f0e3bcf-c987-4ef4-9055-31b947d2807c\") " pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.221884 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.338611 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.436722 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.441557 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content\") pod \"98f5acfe-223f-459b-a995-ae89a695ac32\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.441667 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt8jz\" (UniqueName: \"kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz\") pod \"98f5acfe-223f-459b-a995-ae89a695ac32\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.441701 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities\") pod \"98f5acfe-223f-459b-a995-ae89a695ac32\" (UID: \"98f5acfe-223f-459b-a995-ae89a695ac32\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.442773 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities" (OuterVolumeSpecName: "utilities") pod "98f5acfe-223f-459b-a995-ae89a695ac32" (UID: "98f5acfe-223f-459b-a995-ae89a695ac32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.452133 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.457227 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz" (OuterVolumeSpecName: "kube-api-access-qt8jz") pod "98f5acfe-223f-459b-a995-ae89a695ac32" (UID: "98f5acfe-223f-459b-a995-ae89a695ac32"). InnerVolumeSpecName "kube-api-access-qt8jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.472061 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.499811 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543810 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics\") pod \"41f046c9-8758-4f06-a7d4-8149883734eb\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543855 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca\") pod \"41f046c9-8758-4f06-a7d4-8149883734eb\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543887 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjp7t\" (UniqueName: \"kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t\") pod \"41f046c9-8758-4f06-a7d4-8149883734eb\" (UID: \"41f046c9-8758-4f06-a7d4-8149883734eb\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543952 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities\") pod \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543972 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content\") pod \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.543992 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76t5j\" (UniqueName: \"kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j\") pod \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\" (UID: \"afa6fe4a-b638-4af7-9b26-5d2208d4e185\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.544169 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt8jz\" (UniqueName: \"kubernetes.io/projected/98f5acfe-223f-459b-a995-ae89a695ac32-kube-api-access-qt8jz\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.544201 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.546055 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities" (OuterVolumeSpecName: "utilities") pod "afa6fe4a-b638-4af7-9b26-5d2208d4e185" (UID: "afa6fe4a-b638-4af7-9b26-5d2208d4e185"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.546090 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "41f046c9-8758-4f06-a7d4-8149883734eb" (UID: "41f046c9-8758-4f06-a7d4-8149883734eb"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.546935 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j" (OuterVolumeSpecName: "kube-api-access-76t5j") pod "afa6fe4a-b638-4af7-9b26-5d2208d4e185" (UID: "afa6fe4a-b638-4af7-9b26-5d2208d4e185"). InnerVolumeSpecName "kube-api-access-76t5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.548172 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "41f046c9-8758-4f06-a7d4-8149883734eb" (UID: "41f046c9-8758-4f06-a7d4-8149883734eb"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.552400 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t" (OuterVolumeSpecName: "kube-api-access-cjp7t") pod "41f046c9-8758-4f06-a7d4-8149883734eb" (UID: "41f046c9-8758-4f06-a7d4-8149883734eb"). InnerVolumeSpecName "kube-api-access-cjp7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.565343 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98f5acfe-223f-459b-a995-ae89a695ac32" (UID: "98f5acfe-223f-459b-a995-ae89a695ac32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.615280 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afa6fe4a-b638-4af7-9b26-5d2208d4e185" (UID: "afa6fe4a-b638-4af7-9b26-5d2208d4e185"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645581 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities\") pod \"60d2ac3a-4a78-4bad-af91-633f78900181\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645624 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities\") pod \"9036c1c8-0e08-4924-abee-1d40400d6e96\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645647 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content\") pod \"60d2ac3a-4a78-4bad-af91-633f78900181\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645678 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7czd\" (UniqueName: \"kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd\") pod \"9036c1c8-0e08-4924-abee-1d40400d6e96\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645700 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjpsj\" (UniqueName: \"kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj\") pod \"60d2ac3a-4a78-4bad-af91-633f78900181\" (UID: \"60d2ac3a-4a78-4bad-af91-633f78900181\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645718 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content\") pod \"9036c1c8-0e08-4924-abee-1d40400d6e96\" (UID: \"9036c1c8-0e08-4924-abee-1d40400d6e96\") " Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645906 4993 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645927 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjp7t\" (UniqueName: \"kubernetes.io/projected/41f046c9-8758-4f06-a7d4-8149883734eb-kube-api-access-cjp7t\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645935 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f5acfe-223f-459b-a995-ae89a695ac32-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645945 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645953 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6fe4a-b638-4af7-9b26-5d2208d4e185-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645962 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76t5j\" (UniqueName: \"kubernetes.io/projected/afa6fe4a-b638-4af7-9b26-5d2208d4e185-kube-api-access-76t5j\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.645971 4993 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/41f046c9-8758-4f06-a7d4-8149883734eb-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.649644 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities" (OuterVolumeSpecName: "utilities") pod "60d2ac3a-4a78-4bad-af91-633f78900181" (UID: "60d2ac3a-4a78-4bad-af91-633f78900181"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.649679 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd" (OuterVolumeSpecName: "kube-api-access-c7czd") pod "9036c1c8-0e08-4924-abee-1d40400d6e96" (UID: "9036c1c8-0e08-4924-abee-1d40400d6e96"). InnerVolumeSpecName "kube-api-access-c7czd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.650866 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities" (OuterVolumeSpecName: "utilities") pod "9036c1c8-0e08-4924-abee-1d40400d6e96" (UID: "9036c1c8-0e08-4924-abee-1d40400d6e96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.652608 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj" (OuterVolumeSpecName: "kube-api-access-tjpsj") pod "60d2ac3a-4a78-4bad-af91-633f78900181" (UID: "60d2ac3a-4a78-4bad-af91-633f78900181"). InnerVolumeSpecName "kube-api-access-tjpsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.672011 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9036c1c8-0e08-4924-abee-1d40400d6e96" (UID: "9036c1c8-0e08-4924-abee-1d40400d6e96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.730372 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qvrlt"] Jan 29 12:12:28 crc kubenswrapper[4993]: W0129 12:12:28.732721 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f0e3bcf_c987_4ef4_9055_31b947d2807c.slice/crio-abfa412bebde71a8bb0fb04a3b73650bc3ba4d5b780a168835151ae6660cd97d WatchSource:0}: Error finding container abfa412bebde71a8bb0fb04a3b73650bc3ba4d5b780a168835151ae6660cd97d: Status 404 returned error can't find the container with id abfa412bebde71a8bb0fb04a3b73650bc3ba4d5b780a168835151ae6660cd97d Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.746917 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.746949 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.746959 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7czd\" (UniqueName: \"kubernetes.io/projected/9036c1c8-0e08-4924-abee-1d40400d6e96-kube-api-access-c7czd\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.746970 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjpsj\" (UniqueName: \"kubernetes.io/projected/60d2ac3a-4a78-4bad-af91-633f78900181-kube-api-access-tjpsj\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.746979 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c1c8-0e08-4924-abee-1d40400d6e96-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.802822 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60d2ac3a-4a78-4bad-af91-633f78900181" (UID: "60d2ac3a-4a78-4bad-af91-633f78900181"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.811130 4993 generic.go:334] "Generic (PLEG): container finished" podID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerID="549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5" exitCode=0 Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.811235 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerDied","Data":"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.811263 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlzds" event={"ID":"afa6fe4a-b638-4af7-9b26-5d2208d4e185","Type":"ContainerDied","Data":"4173dd18dd23f980a439777fa1ed539edad0666215f618dac54feb54f0a70630"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.811293 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlzds" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.811326 4993 scope.go:117] "RemoveContainer" containerID="549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.822553 4993 generic.go:334] "Generic (PLEG): container finished" podID="98f5acfe-223f-459b-a995-ae89a695ac32" containerID="5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b" exitCode=0 Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.822619 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzw5p" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.822644 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerDied","Data":"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.822688 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzw5p" event={"ID":"98f5acfe-223f-459b-a995-ae89a695ac32","Type":"ContainerDied","Data":"bc5ff061fe0a5e3fb80564528dd2ed4e2b9da66e251523b7018c007038146ab3"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.831328 4993 generic.go:334] "Generic (PLEG): container finished" podID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerID="4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed" exitCode=0 Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.831407 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n7xrp" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.831428 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerDied","Data":"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.831450 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n7xrp" event={"ID":"9036c1c8-0e08-4924-abee-1d40400d6e96","Type":"ContainerDied","Data":"ec25f9dbb35be43ce5425a9dd7a136c92ee52a22545d828ff89d77eba70ea1af"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.833790 4993 scope.go:117] "RemoveContainer" containerID="573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.834951 4993 generic.go:334] "Generic (PLEG): container finished" podID="60d2ac3a-4a78-4bad-af91-633f78900181" containerID="78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede" exitCode=0 Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.835099 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nh5h" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.835105 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerDied","Data":"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.835152 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nh5h" event={"ID":"60d2ac3a-4a78-4bad-af91-633f78900181","Type":"ContainerDied","Data":"54ac48cebd9efded5c8d3bc2a9eca90b236dbd65f862ec09a07906a40af63a97"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.838843 4993 generic.go:334] "Generic (PLEG): container finished" podID="41f046c9-8758-4f06-a7d4-8149883734eb" containerID="ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f" exitCode=0 Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.838927 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.839133 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerDied","Data":"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.840647 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8k8fn" event={"ID":"41f046c9-8758-4f06-a7d4-8149883734eb","Type":"ContainerDied","Data":"6e83607e406fdb0d871560f73589ef1ccc8ce991b1d26432a497c0f50a84d134"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.843795 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" event={"ID":"2f0e3bcf-c987-4ef4-9055-31b947d2807c","Type":"ContainerStarted","Data":"abfa412bebde71a8bb0fb04a3b73650bc3ba4d5b780a168835151ae6660cd97d"} Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.845386 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.847833 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d2ac3a-4a78-4bad-af91-633f78900181-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.849764 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dlzds"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.864712 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.873450 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zzw5p"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.876711 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.881888 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n7xrp"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.885001 4993 scope.go:117] "RemoveContainer" containerID="a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.892213 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.894885 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8k8fn"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.908907 4993 scope.go:117] "RemoveContainer" containerID="549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.909488 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.909803 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5\": container with ID starting with 549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5 not found: ID does not exist" containerID="549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.909852 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5"} err="failed to get container status \"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5\": rpc error: code = NotFound desc = could not find container \"549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5\": container with ID starting with 549342e0d2aea0f6aa577a9f2d87524a935b72b2ec99383a1c7d191d13f7f4b5 not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.909892 4993 scope.go:117] "RemoveContainer" containerID="573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6" Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.911382 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6\": container with ID starting with 573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6 not found: ID does not exist" containerID="573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.911720 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6"} err="failed to get container status \"573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6\": rpc error: code = NotFound desc = could not find container \"573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6\": container with ID starting with 573e12dd944cac8a42464cb9440db2d364afe6b9b90f0154772ef545301802f6 not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.911785 4993 scope.go:117] "RemoveContainer" containerID="a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f" Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.913977 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f\": container with ID starting with a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f not found: ID does not exist" containerID="a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.914552 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f"} err="failed to get container status \"a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f\": rpc error: code = NotFound desc = could not find container \"a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f\": container with ID starting with a33c41f605da3a898e10a6e2de5a01ce6d71d7ef695b337a7374d9b6340d598f not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.914595 4993 scope.go:117] "RemoveContainer" containerID="5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.914611 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9nh5h"] Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.934442 4993 scope.go:117] "RemoveContainer" containerID="7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.948882 4993 scope.go:117] "RemoveContainer" containerID="4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.967073 4993 scope.go:117] "RemoveContainer" containerID="5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b" Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.967853 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b\": container with ID starting with 5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b not found: ID does not exist" containerID="5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.967914 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b"} err="failed to get container status \"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b\": rpc error: code = NotFound desc = could not find container \"5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b\": container with ID starting with 5136c6bb041bf040a431ceb83a9ad57c70ca1aede918c0103ecf10ba3925556b not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.967952 4993 scope.go:117] "RemoveContainer" containerID="7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6" Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.968568 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6\": container with ID starting with 7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6 not found: ID does not exist" containerID="7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.968633 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6"} err="failed to get container status \"7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6\": rpc error: code = NotFound desc = could not find container \"7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6\": container with ID starting with 7c4f35ca6b0b8b468b0bd96c27a252e56ed6f67f77d66fa54fa6182d4f2fbbf6 not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.968674 4993 scope.go:117] "RemoveContainer" containerID="4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d" Jan 29 12:12:28 crc kubenswrapper[4993]: E0129 12:12:28.969069 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d\": container with ID starting with 4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d not found: ID does not exist" containerID="4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.969096 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d"} err="failed to get container status \"4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d\": rpc error: code = NotFound desc = could not find container \"4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d\": container with ID starting with 4b912147b54f591afb22308d0bd8145392db54da706202c4f84fd4f4a6b0a73d not found: ID does not exist" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.969113 4993 scope.go:117] "RemoveContainer" containerID="4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed" Jan 29 12:12:28 crc kubenswrapper[4993]: I0129 12:12:28.982735 4993 scope.go:117] "RemoveContainer" containerID="c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.002819 4993 scope.go:117] "RemoveContainer" containerID="091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.063941 4993 scope.go:117] "RemoveContainer" containerID="4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.064514 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed\": container with ID starting with 4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed not found: ID does not exist" containerID="4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.064589 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed"} err="failed to get container status \"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed\": rpc error: code = NotFound desc = could not find container \"4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed\": container with ID starting with 4cc1583e835ff9c3632d1db838e3eafb043d5071e49498a030c4246474a400ed not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.064640 4993 scope.go:117] "RemoveContainer" containerID="c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.065236 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0\": container with ID starting with c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0 not found: ID does not exist" containerID="c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.065288 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0"} err="failed to get container status \"c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0\": rpc error: code = NotFound desc = could not find container \"c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0\": container with ID starting with c65308d10c1461f05df8e1546a3aacea0aeceb75cbd6061f6ff4849f06493cd0 not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.065322 4993 scope.go:117] "RemoveContainer" containerID="091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.065629 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5\": container with ID starting with 091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5 not found: ID does not exist" containerID="091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.065666 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5"} err="failed to get container status \"091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5\": rpc error: code = NotFound desc = could not find container \"091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5\": container with ID starting with 091df2d5121e7323a5e2a598b8b42af6386a01569a008aa8460b7d8b3994e5c5 not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.065690 4993 scope.go:117] "RemoveContainer" containerID="78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.080389 4993 scope.go:117] "RemoveContainer" containerID="90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.100124 4993 scope.go:117] "RemoveContainer" containerID="d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.120453 4993 scope.go:117] "RemoveContainer" containerID="78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.120967 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede\": container with ID starting with 78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede not found: ID does not exist" containerID="78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.121008 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede"} err="failed to get container status \"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede\": rpc error: code = NotFound desc = could not find container \"78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede\": container with ID starting with 78da00d0362bb085647bbfcdb9093abc2a2a7001f475295cedaa14c9a9ff0ede not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.121038 4993 scope.go:117] "RemoveContainer" containerID="90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.121444 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2\": container with ID starting with 90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2 not found: ID does not exist" containerID="90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.121484 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2"} err="failed to get container status \"90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2\": rpc error: code = NotFound desc = could not find container \"90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2\": container with ID starting with 90d47eff09cfc529319bf4902aacd57fdc407db848cefc3e379a8e65283968b2 not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.121509 4993 scope.go:117] "RemoveContainer" containerID="d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.122294 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e\": container with ID starting with d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e not found: ID does not exist" containerID="d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.122322 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e"} err="failed to get container status \"d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e\": rpc error: code = NotFound desc = could not find container \"d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e\": container with ID starting with d1422e0d80022ef0aa806959ab94f3e2f03620338e2f7fa5e3c86cc5de62867e not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.122342 4993 scope.go:117] "RemoveContainer" containerID="ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.135240 4993 scope.go:117] "RemoveContainer" containerID="139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.150354 4993 scope.go:117] "RemoveContainer" containerID="ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.150934 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f\": container with ID starting with ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f not found: ID does not exist" containerID="ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.150996 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f"} err="failed to get container status \"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f\": rpc error: code = NotFound desc = could not find container \"ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f\": container with ID starting with ed724b6853f347d50b2c59b227f3108c8de9e0af9cfd0f46a885f34d71689c0f not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.151036 4993 scope.go:117] "RemoveContainer" containerID="139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.151635 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904\": container with ID starting with 139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904 not found: ID does not exist" containerID="139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.151682 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904"} err="failed to get container status \"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904\": rpc error: code = NotFound desc = could not find container \"139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904\": container with ID starting with 139ed8b232085ffbf61d36347483802092d2b7b23e051a5ae816a89103036904 not found: ID does not exist" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.196522 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" path="/var/lib/kubelet/pods/41f046c9-8758-4f06-a7d4-8149883734eb/volumes" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.197026 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" path="/var/lib/kubelet/pods/60d2ac3a-4a78-4bad-af91-633f78900181/volumes" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.197611 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" path="/var/lib/kubelet/pods/9036c1c8-0e08-4924-abee-1d40400d6e96/volumes" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.198597 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" path="/var/lib/kubelet/pods/98f5acfe-223f-459b-a995-ae89a695ac32/volumes" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.199279 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" path="/var/lib/kubelet/pods/afa6fe4a-b638-4af7-9b26-5d2208d4e185/volumes" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675618 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jbcpr"] Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675845 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675861 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675875 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675894 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675905 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675913 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675925 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675932 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675941 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675948 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675959 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675967 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675976 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675983 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.675992 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.675999 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676010 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676017 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676029 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676051 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="extract-utilities" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676063 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676070 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676080 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676087 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676094 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676100 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="extract-content" Jan 29 12:12:29 crc kubenswrapper[4993]: E0129 12:12:29.676109 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676118 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676232 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="9036c1c8-0e08-4924-abee-1d40400d6e96" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676246 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676262 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f5acfe-223f-459b-a995-ae89a695ac32" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676272 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa6fe4a-b638-4af7-9b26-5d2208d4e185" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676283 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d2ac3a-4a78-4bad-af91-633f78900181" containerName="registry-server" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.676460 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f046c9-8758-4f06-a7d4-8149883734eb" containerName="marketplace-operator" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.677069 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.679108 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.692504 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbcpr"] Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.852083 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" event={"ID":"2f0e3bcf-c987-4ef4-9055-31b947d2807c","Type":"ContainerStarted","Data":"9fb673bbe1ef43cf6bcba8eb0b8acec3044afd811045db9b233c533876a85b31"} Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.852518 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.857661 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.860217 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-utilities\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.860323 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-catalog-content\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.860394 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfrv7\" (UniqueName: \"kubernetes.io/projected/a175a8e3-3022-4953-9419-4fc74c9fb0e0-kube-api-access-cfrv7\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.872604 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qvrlt" podStartSLOduration=2.872582167 podStartE2EDuration="2.872582167s" podCreationTimestamp="2026-01-29 12:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:12:29.869074169 +0000 UTC m=+433.884204295" watchObservedRunningTime="2026-01-29 12:12:29.872582167 +0000 UTC m=+433.887712313" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.961779 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-catalog-content\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.961828 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfrv7\" (UniqueName: \"kubernetes.io/projected/a175a8e3-3022-4953-9419-4fc74c9fb0e0-kube-api-access-cfrv7\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.961933 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-utilities\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.963151 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-utilities\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.963269 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a175a8e3-3022-4953-9419-4fc74c9fb0e0-catalog-content\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.983977 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfrv7\" (UniqueName: \"kubernetes.io/projected/a175a8e3-3022-4953-9419-4fc74c9fb0e0-kube-api-access-cfrv7\") pod \"certified-operators-jbcpr\" (UID: \"a175a8e3-3022-4953-9419-4fc74c9fb0e0\") " pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:29 crc kubenswrapper[4993]: I0129 12:12:29.994611 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.391641 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbcpr"] Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.662894 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sjktr"] Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.663961 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.669010 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.717587 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjktr"] Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.774827 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-catalog-content\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.774919 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-utilities\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.775018 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh7mw\" (UniqueName: \"kubernetes.io/projected/768166e9-0d56-4f39-9901-0da8ea9bd051-kube-api-access-lh7mw\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.868549 4993 generic.go:334] "Generic (PLEG): container finished" podID="a175a8e3-3022-4953-9419-4fc74c9fb0e0" containerID="d93197726002097c3f06f93e61d14703c7942c1bb46a4ba3dad5164c5c4ef0ea" exitCode=0 Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.869995 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbcpr" event={"ID":"a175a8e3-3022-4953-9419-4fc74c9fb0e0","Type":"ContainerDied","Data":"d93197726002097c3f06f93e61d14703c7942c1bb46a4ba3dad5164c5c4ef0ea"} Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.870037 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbcpr" event={"ID":"a175a8e3-3022-4953-9419-4fc74c9fb0e0","Type":"ContainerStarted","Data":"959bcefb1dc8ce419c231d0874da42abdb62e81e379f79656d2d96b3c9930c88"} Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.877067 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-utilities\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.877432 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-utilities\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.877547 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh7mw\" (UniqueName: \"kubernetes.io/projected/768166e9-0d56-4f39-9901-0da8ea9bd051-kube-api-access-lh7mw\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.877705 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-catalog-content\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.878094 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/768166e9-0d56-4f39-9901-0da8ea9bd051-catalog-content\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:30 crc kubenswrapper[4993]: I0129 12:12:30.898577 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh7mw\" (UniqueName: \"kubernetes.io/projected/768166e9-0d56-4f39-9901-0da8ea9bd051-kube-api-access-lh7mw\") pod \"redhat-marketplace-sjktr\" (UID: \"768166e9-0d56-4f39-9901-0da8ea9bd051\") " pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.018908 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.404625 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sjktr"] Jan 29 12:12:31 crc kubenswrapper[4993]: W0129 12:12:31.408278 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod768166e9_0d56_4f39_9901_0da8ea9bd051.slice/crio-202d9b19dcac96e055e81102d9aae0be28d5ba2f0fcbb4d0fbf3f6d8e299032a WatchSource:0}: Error finding container 202d9b19dcac96e055e81102d9aae0be28d5ba2f0fcbb4d0fbf3f6d8e299032a: Status 404 returned error can't find the container with id 202d9b19dcac96e055e81102d9aae0be28d5ba2f0fcbb4d0fbf3f6d8e299032a Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.875357 4993 generic.go:334] "Generic (PLEG): container finished" podID="a175a8e3-3022-4953-9419-4fc74c9fb0e0" containerID="1f1802177230457b093ae9bad82f451bb246edcfcde7b8acc27abc7f3a22c6fe" exitCode=0 Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.875438 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbcpr" event={"ID":"a175a8e3-3022-4953-9419-4fc74c9fb0e0","Type":"ContainerDied","Data":"1f1802177230457b093ae9bad82f451bb246edcfcde7b8acc27abc7f3a22c6fe"} Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.877482 4993 generic.go:334] "Generic (PLEG): container finished" podID="768166e9-0d56-4f39-9901-0da8ea9bd051" containerID="150293314a513df3462cc594d6d7b41c3dab8709123ca6775f1973ce815d4a5c" exitCode=0 Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.877597 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjktr" event={"ID":"768166e9-0d56-4f39-9901-0da8ea9bd051","Type":"ContainerDied","Data":"150293314a513df3462cc594d6d7b41c3dab8709123ca6775f1973ce815d4a5c"} Jan 29 12:12:31 crc kubenswrapper[4993]: I0129 12:12:31.877643 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjktr" event={"ID":"768166e9-0d56-4f39-9901-0da8ea9bd051","Type":"ContainerStarted","Data":"202d9b19dcac96e055e81102d9aae0be28d5ba2f0fcbb4d0fbf3f6d8e299032a"} Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.074153 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fpf2p"] Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.075771 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.078076 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.082059 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fpf2p"] Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.192600 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrdv2\" (UniqueName: \"kubernetes.io/projected/f36b3ae1-b711-44a3-900e-2c51754a7109-kube-api-access-rrdv2\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.192673 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-utilities\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.192825 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-catalog-content\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.294043 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrdv2\" (UniqueName: \"kubernetes.io/projected/f36b3ae1-b711-44a3-900e-2c51754a7109-kube-api-access-rrdv2\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.294123 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-utilities\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.294165 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-catalog-content\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.294758 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-catalog-content\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.294755 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f36b3ae1-b711-44a3-900e-2c51754a7109-utilities\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.311415 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrdv2\" (UniqueName: \"kubernetes.io/projected/f36b3ae1-b711-44a3-900e-2c51754a7109-kube-api-access-rrdv2\") pod \"redhat-operators-fpf2p\" (UID: \"f36b3ae1-b711-44a3-900e-2c51754a7109\") " pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.388556 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.792669 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fpf2p"] Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.887298 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbcpr" event={"ID":"a175a8e3-3022-4953-9419-4fc74c9fb0e0","Type":"ContainerStarted","Data":"4d77ff3c11c73cef95d55d5ab67857a3326f4050e1795a7ec1bae7cb4a82ec09"} Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.888634 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpf2p" event={"ID":"f36b3ae1-b711-44a3-900e-2c51754a7109","Type":"ContainerStarted","Data":"520081d537277f156d73eeae1b629361bf7cb39554de65f2f4cfdf47feb2893e"} Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.890753 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjktr" event={"ID":"768166e9-0d56-4f39-9901-0da8ea9bd051","Type":"ContainerStarted","Data":"9c89cdac598ea06dd04afc3a11b6b3dd1f8ea09b6e8a7c519e5e12104faba0fe"} Jan 29 12:12:32 crc kubenswrapper[4993]: I0129 12:12:32.913013 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jbcpr" podStartSLOduration=2.4490172599999998 podStartE2EDuration="3.912995951s" podCreationTimestamp="2026-01-29 12:12:29 +0000 UTC" firstStartedPulling="2026-01-29 12:12:30.870666211 +0000 UTC m=+434.885796337" lastFinishedPulling="2026-01-29 12:12:32.334644902 +0000 UTC m=+436.349775028" observedRunningTime="2026-01-29 12:12:32.906153489 +0000 UTC m=+436.921283635" watchObservedRunningTime="2026-01-29 12:12:32.912995951 +0000 UTC m=+436.928126077" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.062543 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9w5fq"] Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.063847 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.066602 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.087042 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9w5fq"] Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.211580 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-utilities\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.211641 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5js9\" (UniqueName: \"kubernetes.io/projected/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-kube-api-access-x5js9\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.211677 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-catalog-content\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.313029 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-utilities\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.313122 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5js9\" (UniqueName: \"kubernetes.io/projected/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-kube-api-access-x5js9\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.313201 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-catalog-content\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.313768 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-catalog-content\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.314067 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-utilities\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.336286 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5js9\" (UniqueName: \"kubernetes.io/projected/98b3b4b4-59d6-4e70-ab5b-f763edbc691a-kube-api-access-x5js9\") pod \"community-operators-9w5fq\" (UID: \"98b3b4b4-59d6-4e70-ab5b-f763edbc691a\") " pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.380124 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.797089 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9w5fq"] Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.896046 4993 generic.go:334] "Generic (PLEG): container finished" podID="768166e9-0d56-4f39-9901-0da8ea9bd051" containerID="9c89cdac598ea06dd04afc3a11b6b3dd1f8ea09b6e8a7c519e5e12104faba0fe" exitCode=0 Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.896129 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjktr" event={"ID":"768166e9-0d56-4f39-9901-0da8ea9bd051","Type":"ContainerDied","Data":"9c89cdac598ea06dd04afc3a11b6b3dd1f8ea09b6e8a7c519e5e12104faba0fe"} Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.898230 4993 generic.go:334] "Generic (PLEG): container finished" podID="f36b3ae1-b711-44a3-900e-2c51754a7109" containerID="3eee0babc506c9572603e4b8fedfd08e9b93d6d58e57f60c3bcb8c51e13b5f52" exitCode=0 Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.898302 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpf2p" event={"ID":"f36b3ae1-b711-44a3-900e-2c51754a7109","Type":"ContainerDied","Data":"3eee0babc506c9572603e4b8fedfd08e9b93d6d58e57f60c3bcb8c51e13b5f52"} Jan 29 12:12:33 crc kubenswrapper[4993]: I0129 12:12:33.900152 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w5fq" event={"ID":"98b3b4b4-59d6-4e70-ab5b-f763edbc691a","Type":"ContainerStarted","Data":"99543fefc26e1b4f4cccb4af84d60e6f768efab881f2521bd828c95e366a9b74"} Jan 29 12:12:34 crc kubenswrapper[4993]: I0129 12:12:34.908494 4993 generic.go:334] "Generic (PLEG): container finished" podID="98b3b4b4-59d6-4e70-ab5b-f763edbc691a" containerID="d58456979586799ddd6da3fc7e12b1bd7fbbedd4de81d6fb71a2d2ae4f91911f" exitCode=0 Jan 29 12:12:34 crc kubenswrapper[4993]: I0129 12:12:34.908588 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w5fq" event={"ID":"98b3b4b4-59d6-4e70-ab5b-f763edbc691a","Type":"ContainerDied","Data":"d58456979586799ddd6da3fc7e12b1bd7fbbedd4de81d6fb71a2d2ae4f91911f"} Jan 29 12:12:34 crc kubenswrapper[4993]: I0129 12:12:34.911855 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sjktr" event={"ID":"768166e9-0d56-4f39-9901-0da8ea9bd051","Type":"ContainerStarted","Data":"b1c83fec8c26550619a2812180809128eb782482c9c2def4430f514dc8d2d397"} Jan 29 12:12:34 crc kubenswrapper[4993]: I0129 12:12:34.949320 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sjktr" podStartSLOduration=2.525899399 podStartE2EDuration="4.949302382s" podCreationTimestamp="2026-01-29 12:12:30 +0000 UTC" firstStartedPulling="2026-01-29 12:12:31.878845207 +0000 UTC m=+435.893975323" lastFinishedPulling="2026-01-29 12:12:34.30224818 +0000 UTC m=+438.317378306" observedRunningTime="2026-01-29 12:12:34.948408906 +0000 UTC m=+438.963539052" watchObservedRunningTime="2026-01-29 12:12:34.949302382 +0000 UTC m=+438.964432498" Jan 29 12:12:35 crc kubenswrapper[4993]: I0129 12:12:35.918606 4993 generic.go:334] "Generic (PLEG): container finished" podID="f36b3ae1-b711-44a3-900e-2c51754a7109" containerID="ca140fa2d7dbecafc3b850f40a470092df52d3eda317d3d8952bc2b3c50fd026" exitCode=0 Jan 29 12:12:35 crc kubenswrapper[4993]: I0129 12:12:35.918666 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpf2p" event={"ID":"f36b3ae1-b711-44a3-900e-2c51754a7109","Type":"ContainerDied","Data":"ca140fa2d7dbecafc3b850f40a470092df52d3eda317d3d8952bc2b3c50fd026"} Jan 29 12:12:35 crc kubenswrapper[4993]: I0129 12:12:35.921158 4993 generic.go:334] "Generic (PLEG): container finished" podID="98b3b4b4-59d6-4e70-ab5b-f763edbc691a" containerID="7c1c62f5aac855ca8e3285da4aa9f94480188e26457c83076a0addd1c4794fbd" exitCode=0 Jan 29 12:12:35 crc kubenswrapper[4993]: I0129 12:12:35.922251 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w5fq" event={"ID":"98b3b4b4-59d6-4e70-ab5b-f763edbc691a","Type":"ContainerDied","Data":"7c1c62f5aac855ca8e3285da4aa9f94480188e26457c83076a0addd1c4794fbd"} Jan 29 12:12:36 crc kubenswrapper[4993]: I0129 12:12:36.928588 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpf2p" event={"ID":"f36b3ae1-b711-44a3-900e-2c51754a7109","Type":"ContainerStarted","Data":"c91bc639ea6cada40edf40eb19970fd299ddd53279dfa69e485ed3bf9c6fb2af"} Jan 29 12:12:36 crc kubenswrapper[4993]: I0129 12:12:36.931390 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w5fq" event={"ID":"98b3b4b4-59d6-4e70-ab5b-f763edbc691a","Type":"ContainerStarted","Data":"f08e9ae71418c28cf3fc327f862330143e469af43cac9122f3ce9ab2b535e8a6"} Jan 29 12:12:36 crc kubenswrapper[4993]: I0129 12:12:36.952301 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fpf2p" podStartSLOduration=2.477065165 podStartE2EDuration="4.952279809s" podCreationTimestamp="2026-01-29 12:12:32 +0000 UTC" firstStartedPulling="2026-01-29 12:12:33.899735136 +0000 UTC m=+437.914865262" lastFinishedPulling="2026-01-29 12:12:36.37494978 +0000 UTC m=+440.390079906" observedRunningTime="2026-01-29 12:12:36.950555171 +0000 UTC m=+440.965685297" watchObservedRunningTime="2026-01-29 12:12:36.952279809 +0000 UTC m=+440.967409935" Jan 29 12:12:36 crc kubenswrapper[4993]: I0129 12:12:36.975474 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9w5fq" podStartSLOduration=2.58945667 podStartE2EDuration="3.975454308s" podCreationTimestamp="2026-01-29 12:12:33 +0000 UTC" firstStartedPulling="2026-01-29 12:12:34.915095403 +0000 UTC m=+438.930225529" lastFinishedPulling="2026-01-29 12:12:36.301093041 +0000 UTC m=+440.316223167" observedRunningTime="2026-01-29 12:12:36.970737847 +0000 UTC m=+440.985867973" watchObservedRunningTime="2026-01-29 12:12:36.975454308 +0000 UTC m=+440.990584444" Jan 29 12:12:39 crc kubenswrapper[4993]: I0129 12:12:39.995072 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:39 crc kubenswrapper[4993]: I0129 12:12:39.995818 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:40 crc kubenswrapper[4993]: I0129 12:12:40.050647 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:40 crc kubenswrapper[4993]: I0129 12:12:40.994330 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jbcpr" Jan 29 12:12:41 crc kubenswrapper[4993]: I0129 12:12:41.019175 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:41 crc kubenswrapper[4993]: I0129 12:12:41.019242 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:41 crc kubenswrapper[4993]: I0129 12:12:41.057277 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:42 crc kubenswrapper[4993]: I0129 12:12:42.005830 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sjktr" Jan 29 12:12:42 crc kubenswrapper[4993]: I0129 12:12:42.389565 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:42 crc kubenswrapper[4993]: I0129 12:12:42.389664 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:42 crc kubenswrapper[4993]: I0129 12:12:42.447011 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:43 crc kubenswrapper[4993]: I0129 12:12:43.009879 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fpf2p" Jan 29 12:12:43 crc kubenswrapper[4993]: I0129 12:12:43.380875 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:43 crc kubenswrapper[4993]: I0129 12:12:43.381245 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:43 crc kubenswrapper[4993]: I0129 12:12:43.416609 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:12:44 crc kubenswrapper[4993]: I0129 12:12:44.019749 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9w5fq" Jan 29 12:14:17 crc kubenswrapper[4993]: I0129 12:14:17.435365 4993 scope.go:117] "RemoveContainer" containerID="23bc79e9d11a23d95c564334222f47db12214b3607bfa71c67e6b7bdb2bf00d2" Jan 29 12:14:18 crc kubenswrapper[4993]: I0129 12:14:18.155705 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:14:18 crc kubenswrapper[4993]: I0129 12:14:18.155798 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:14:48 crc kubenswrapper[4993]: I0129 12:14:48.156447 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:14:48 crc kubenswrapper[4993]: I0129 12:14:48.157074 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.175518 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49"] Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.177616 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.179558 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.181345 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.192593 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49"] Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.297966 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.298031 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmrrq\" (UniqueName: \"kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.298111 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.399701 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.399778 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmrrq\" (UniqueName: \"kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.399814 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.400927 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.405914 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.424064 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmrrq\" (UniqueName: \"kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq\") pod \"collect-profiles-29494815-2xw49\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.501115 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:00 crc kubenswrapper[4993]: I0129 12:15:00.955498 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49"] Jan 29 12:15:01 crc kubenswrapper[4993]: I0129 12:15:01.765606 4993 generic.go:334] "Generic (PLEG): container finished" podID="d48e225a-e472-497b-b1af-08f427868ed8" containerID="d257f61af223ff42f9577780254ceb22e7311ca2c78492fdefdc2540d941ad71" exitCode=0 Jan 29 12:15:01 crc kubenswrapper[4993]: I0129 12:15:01.765701 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" event={"ID":"d48e225a-e472-497b-b1af-08f427868ed8","Type":"ContainerDied","Data":"d257f61af223ff42f9577780254ceb22e7311ca2c78492fdefdc2540d941ad71"} Jan 29 12:15:01 crc kubenswrapper[4993]: I0129 12:15:01.766034 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" event={"ID":"d48e225a-e472-497b-b1af-08f427868ed8","Type":"ContainerStarted","Data":"059ec57d5c602cfd528e9846f20579e254dd255ecbc3e44f65a44dbf1842b79e"} Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.545404 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.679611 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume\") pod \"d48e225a-e472-497b-b1af-08f427868ed8\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.679701 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume\") pod \"d48e225a-e472-497b-b1af-08f427868ed8\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.679789 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmrrq\" (UniqueName: \"kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq\") pod \"d48e225a-e472-497b-b1af-08f427868ed8\" (UID: \"d48e225a-e472-497b-b1af-08f427868ed8\") " Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.680617 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume" (OuterVolumeSpecName: "config-volume") pod "d48e225a-e472-497b-b1af-08f427868ed8" (UID: "d48e225a-e472-497b-b1af-08f427868ed8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.685411 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq" (OuterVolumeSpecName: "kube-api-access-vmrrq") pod "d48e225a-e472-497b-b1af-08f427868ed8" (UID: "d48e225a-e472-497b-b1af-08f427868ed8"). InnerVolumeSpecName "kube-api-access-vmrrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.686337 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d48e225a-e472-497b-b1af-08f427868ed8" (UID: "d48e225a-e472-497b-b1af-08f427868ed8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.777432 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" event={"ID":"d48e225a-e472-497b-b1af-08f427868ed8","Type":"ContainerDied","Data":"059ec57d5c602cfd528e9846f20579e254dd255ecbc3e44f65a44dbf1842b79e"} Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.777471 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="059ec57d5c602cfd528e9846f20579e254dd255ecbc3e44f65a44dbf1842b79e" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.777570 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.781345 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmrrq\" (UniqueName: \"kubernetes.io/projected/d48e225a-e472-497b-b1af-08f427868ed8-kube-api-access-vmrrq\") on node \"crc\" DevicePath \"\"" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.781454 4993 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d48e225a-e472-497b-b1af-08f427868ed8-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:15:03 crc kubenswrapper[4993]: I0129 12:15:03.781532 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d48e225a-e472-497b-b1af-08f427868ed8-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:15:17 crc kubenswrapper[4993]: I0129 12:15:17.480903 4993 scope.go:117] "RemoveContainer" containerID="3904233c3a323ecbf0b9d65c3d54252e4299a85773b27eca90f1b6d12eaf4298" Jan 29 12:15:17 crc kubenswrapper[4993]: I0129 12:15:17.505264 4993 scope.go:117] "RemoveContainer" containerID="7ad2efeeca225dc2763a2b1b03b14ec0cd497d735f2c5b3f59a3854732aa3d84" Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.155685 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.156076 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.156376 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.157414 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.157641 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b" gracePeriod=600 Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.873022 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b" exitCode=0 Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.873099 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b"} Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.873581 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd"} Jan 29 12:15:18 crc kubenswrapper[4993]: I0129 12:15:18.873615 4993 scope.go:117] "RemoveContainer" containerID="da745a0101c0520bfecd95c6f64cac6e8a724e36c919a52f1da8013b94dc3c89" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.057254 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j4lz6"] Jan 29 12:16:03 crc kubenswrapper[4993]: E0129 12:16:03.058065 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48e225a-e472-497b-b1af-08f427868ed8" containerName="collect-profiles" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.058081 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48e225a-e472-497b-b1af-08f427868ed8" containerName="collect-profiles" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.058221 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48e225a-e472-497b-b1af-08f427868ed8" containerName="collect-profiles" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.058721 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.073872 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j4lz6"] Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.175587 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-trusted-ca\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.175724 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-tls\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.175829 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-bound-sa-token\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.175934 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a4098744-de19-4e59-98c1-57ce7c0a9ac1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.175998 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdxnn\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-kube-api-access-xdxnn\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.176134 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-certificates\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.176313 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a4098744-de19-4e59-98c1-57ce7c0a9ac1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.176379 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.202140 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277094 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a4098744-de19-4e59-98c1-57ce7c0a9ac1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277203 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-trusted-ca\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277250 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-tls\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277283 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-bound-sa-token\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277314 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a4098744-de19-4e59-98c1-57ce7c0a9ac1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277371 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdxnn\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-kube-api-access-xdxnn\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277408 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-certificates\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.277765 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a4098744-de19-4e59-98c1-57ce7c0a9ac1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.279004 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-certificates\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.281317 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4098744-de19-4e59-98c1-57ce7c0a9ac1-trusted-ca\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.283343 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-registry-tls\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.285384 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a4098744-de19-4e59-98c1-57ce7c0a9ac1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.296001 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdxnn\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-kube-api-access-xdxnn\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.301321 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4098744-de19-4e59-98c1-57ce7c0a9ac1-bound-sa-token\") pod \"image-registry-66df7c8f76-j4lz6\" (UID: \"a4098744-de19-4e59-98c1-57ce7c0a9ac1\") " pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.375479 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:03 crc kubenswrapper[4993]: I0129 12:16:03.567246 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j4lz6"] Jan 29 12:16:04 crc kubenswrapper[4993]: I0129 12:16:04.162084 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" event={"ID":"a4098744-de19-4e59-98c1-57ce7c0a9ac1","Type":"ContainerStarted","Data":"94546a4bb91e72225c222ff8eb8d6856871ff0f7b4032c4d34b4de8c29469caa"} Jan 29 12:16:04 crc kubenswrapper[4993]: I0129 12:16:04.162478 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:04 crc kubenswrapper[4993]: I0129 12:16:04.162493 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" event={"ID":"a4098744-de19-4e59-98c1-57ce7c0a9ac1","Type":"ContainerStarted","Data":"15e45d4a99664cb18c838eac3cb91ae5a58912c221cf6696d0f8bd885faa748d"} Jan 29 12:16:04 crc kubenswrapper[4993]: I0129 12:16:04.187621 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" podStartSLOduration=1.187600656 podStartE2EDuration="1.187600656s" podCreationTimestamp="2026-01-29 12:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:16:04.187379611 +0000 UTC m=+648.202509757" watchObservedRunningTime="2026-01-29 12:16:04.187600656 +0000 UTC m=+648.202730782" Jan 29 12:16:23 crc kubenswrapper[4993]: I0129 12:16:23.382213 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-j4lz6" Jan 29 12:16:23 crc kubenswrapper[4993]: I0129 12:16:23.436405 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:16:48 crc kubenswrapper[4993]: I0129 12:16:48.486502 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" podUID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" containerName="registry" containerID="cri-o://2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417" gracePeriod=30 Jan 29 12:16:48 crc kubenswrapper[4993]: I0129 12:16:48.869039 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.059686 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.059800 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.059850 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.059912 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.060143 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.061355 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-874q9\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.061391 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.061579 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.061632 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca\") pod \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\" (UID: \"0e93bec6-c6d1-4018-8da3-73f918e4d5b6\") " Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.062766 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.063255 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.063360 4993 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.067447 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.068089 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.068791 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.068961 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9" (OuterVolumeSpecName: "kube-api-access-874q9") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "kube-api-access-874q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.083997 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.092876 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "0e93bec6-c6d1-4018-8da3-73f918e4d5b6" (UID: "0e93bec6-c6d1-4018-8da3-73f918e4d5b6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.164578 4993 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.164928 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-874q9\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-kube-api-access-874q9\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.164946 4993 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.164960 4993 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.164972 4993 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e93bec6-c6d1-4018-8da3-73f918e4d5b6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.504586 4993 generic.go:334] "Generic (PLEG): container finished" podID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" containerID="2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417" exitCode=0 Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.504636 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" event={"ID":"0e93bec6-c6d1-4018-8da3-73f918e4d5b6","Type":"ContainerDied","Data":"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417"} Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.504669 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" event={"ID":"0e93bec6-c6d1-4018-8da3-73f918e4d5b6","Type":"ContainerDied","Data":"6376f74ddec262b10a1930ceb69be03cde928f084c7de903a4b2c4a52474bb3f"} Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.504690 4993 scope.go:117] "RemoveContainer" containerID="2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.504810 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jz5vk" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.526601 4993 scope.go:117] "RemoveContainer" containerID="2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417" Jan 29 12:16:49 crc kubenswrapper[4993]: E0129 12:16:49.527441 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417\": container with ID starting with 2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417 not found: ID does not exist" containerID="2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.527483 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417"} err="failed to get container status \"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417\": rpc error: code = NotFound desc = could not find container \"2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417\": container with ID starting with 2d269bb4f8f1194d73081ad41156a58cd96372b01546d99a47388c398ec01417 not found: ID does not exist" Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.535423 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:16:49 crc kubenswrapper[4993]: I0129 12:16:49.541598 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jz5vk"] Jan 29 12:16:51 crc kubenswrapper[4993]: I0129 12:16:51.197400 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" path="/var/lib/kubelet/pods/0e93bec6-c6d1-4018-8da3-73f918e4d5b6/volumes" Jan 29 12:17:18 crc kubenswrapper[4993]: I0129 12:17:18.156462 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:17:18 crc kubenswrapper[4993]: I0129 12:17:18.158296 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:17:48 crc kubenswrapper[4993]: I0129 12:17:48.157370 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:17:48 crc kubenswrapper[4993]: I0129 12:17:48.158056 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:17:53 crc kubenswrapper[4993]: I0129 12:17:53.823884 4993 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.156541 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.157167 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.157255 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.157922 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.157987 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd" gracePeriod=600 Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.996875 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd" exitCode=0 Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.996922 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd"} Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.997484 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141"} Jan 29 12:18:18 crc kubenswrapper[4993]: I0129 12:18:18.997512 4993 scope.go:117] "RemoveContainer" containerID="79a7ea33d01ed9009bf34e364340b4116eda5cee9514868a7fd26a32ec6c826b" Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.932967 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzr6r"] Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.933789 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-controller" containerID="cri-o://771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934169 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="sbdb" containerID="cri-o://46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934226 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="nbdb" containerID="cri-o://00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934260 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="northd" containerID="cri-o://f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934291 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934319 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-node" containerID="cri-o://19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.934344 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-acl-logging" containerID="cri-o://119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" gracePeriod=30 Jan 29 12:19:04 crc kubenswrapper[4993]: I0129 12:19:04.970097 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" containerID="cri-o://b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" gracePeriod=30 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.282137 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/3.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.284331 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovn-acl-logging/0.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.284873 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovn-controller/0.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.285367 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.287967 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovnkube-controller/3.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290014 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovn-acl-logging/0.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290567 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzr6r_c6387d35-c6e4-42d4-9f89-6f2940878f8a/ovn-controller/0.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290940 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290968 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290980 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290998 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291012 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291021 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" exitCode=0 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291029 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" exitCode=143 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291038 4993 generic.go:334] "Generic (PLEG): container finished" podID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" exitCode=143 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291035 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.290974 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291237 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291253 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291262 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291288 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291422 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291435 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291445 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291454 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291461 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291466 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291471 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291476 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291481 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291486 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291490 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291498 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291509 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291522 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291566 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291578 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291583 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291588 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291593 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291598 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291623 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291630 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291640 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291653 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291707 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291714 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291721 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291727 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291732 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291737 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291743 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291748 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291753 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291762 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzr6r" event={"ID":"c6387d35-c6e4-42d4-9f89-6f2940878f8a","Type":"ContainerDied","Data":"225889fd3889840ab1c01c1dd9a0aa1ad8002f0e6a426e098d662f7517f0732a"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291773 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291779 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291784 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291789 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291794 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291799 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291803 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291808 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291813 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.291817 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.293212 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/2.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.293630 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/1.log" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.293668 4993 generic.go:334] "Generic (PLEG): container finished" podID="d22ea793-b0f9-4b2c-8ea7-ed61728d900f" containerID="6d31403f624f5e15e97a01b0394e0dfbf2c956eebd89eee0e1fd00dcfde615b6" exitCode=2 Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.293691 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerDied","Data":"6d31403f624f5e15e97a01b0394e0dfbf2c956eebd89eee0e1fd00dcfde615b6"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.293749 4993 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b"} Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.294151 4993 scope.go:117] "RemoveContainer" containerID="6d31403f624f5e15e97a01b0394e0dfbf2c956eebd89eee0e1fd00dcfde615b6" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.316469 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.336920 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mx5jx"] Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337122 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337134 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337145 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kubecfg-setup" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337151 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kubecfg-setup" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337159 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="northd" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337166 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="northd" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337173 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="sbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337179 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="sbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337205 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337211 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337217 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" containerName="registry" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337223 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" containerName="registry" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337234 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-node" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337239 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-node" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337247 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337253 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337260 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337266 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337272 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337278 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337284 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337291 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337298 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="nbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337303 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="nbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337313 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-acl-logging" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337319 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-acl-logging" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337403 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337410 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337416 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-acl-logging" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337424 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-node" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337432 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="northd" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337442 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e93bec6-c6d1-4018-8da3-73f918e4d5b6" containerName="registry" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337449 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337456 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337462 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovn-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337468 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="nbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337474 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="sbdb" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.337553 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337560 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.337653 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.338021 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" containerName="ovnkube-controller" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.343264 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.360335 4993 scope.go:117] "RemoveContainer" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.381527 4993 scope.go:117] "RemoveContainer" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.395253 4993 scope.go:117] "RemoveContainer" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408152 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408239 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408261 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408290 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408318 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408337 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408372 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408391 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408415 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t8gs\" (UniqueName: \"kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408442 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408471 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408492 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408531 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408547 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408563 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408584 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408598 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408611 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408625 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408643 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\" (UID: \"c6387d35-c6e4-42d4-9f89-6f2940878f8a\") " Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408402 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409354 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409381 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket" (OuterVolumeSpecName: "log-socket") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409655 4993 scope.go:117] "RemoveContainer" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409654 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log" (OuterVolumeSpecName: "node-log") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409148 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408446 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408507 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408524 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408543 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408568 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408871 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.408888 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409040 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409055 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash" (OuterVolumeSpecName: "host-slash") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409067 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409079 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.409146 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.414706 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs" (OuterVolumeSpecName: "kube-api-access-5t8gs") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "kube-api-access-5t8gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.414862 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.422998 4993 scope.go:117] "RemoveContainer" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.423043 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c6387d35-c6e4-42d4-9f89-6f2940878f8a" (UID: "c6387d35-c6e4-42d4-9f89-6f2940878f8a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.435483 4993 scope.go:117] "RemoveContainer" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.448870 4993 scope.go:117] "RemoveContainer" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.466411 4993 scope.go:117] "RemoveContainer" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.480999 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.481483 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.481525 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} err="failed to get container status \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.481550 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.482066 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": container with ID starting with 08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef not found: ID does not exist" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.482101 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} err="failed to get container status \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": rpc error: code = NotFound desc = could not find container \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": container with ID starting with 08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.482127 4993 scope.go:117] "RemoveContainer" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.482551 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": container with ID starting with 46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828 not found: ID does not exist" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.482607 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} err="failed to get container status \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": rpc error: code = NotFound desc = could not find container \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": container with ID starting with 46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.482639 4993 scope.go:117] "RemoveContainer" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.482999 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": container with ID starting with 00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc not found: ID does not exist" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483021 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} err="failed to get container status \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": rpc error: code = NotFound desc = could not find container \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": container with ID starting with 00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483036 4993 scope.go:117] "RemoveContainer" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.483303 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": container with ID starting with f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa not found: ID does not exist" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483334 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} err="failed to get container status \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": rpc error: code = NotFound desc = could not find container \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": container with ID starting with f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483353 4993 scope.go:117] "RemoveContainer" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.483603 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": container with ID starting with 16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee not found: ID does not exist" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483631 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} err="failed to get container status \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": rpc error: code = NotFound desc = could not find container \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": container with ID starting with 16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483654 4993 scope.go:117] "RemoveContainer" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.483956 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": container with ID starting with 19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e not found: ID does not exist" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.483984 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} err="failed to get container status \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": rpc error: code = NotFound desc = could not find container \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": container with ID starting with 19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.484002 4993 scope.go:117] "RemoveContainer" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.484460 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": container with ID starting with 119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18 not found: ID does not exist" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.484486 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} err="failed to get container status \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": rpc error: code = NotFound desc = could not find container \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": container with ID starting with 119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.484505 4993 scope.go:117] "RemoveContainer" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.484931 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": container with ID starting with 771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7 not found: ID does not exist" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.484956 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} err="failed to get container status \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": rpc error: code = NotFound desc = could not find container \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": container with ID starting with 771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.484972 4993 scope.go:117] "RemoveContainer" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: E0129 12:19:05.485278 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": container with ID starting with 0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c not found: ID does not exist" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485310 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} err="failed to get container status \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": rpc error: code = NotFound desc = could not find container \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": container with ID starting with 0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485347 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485588 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} err="failed to get container status \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485613 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485855 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} err="failed to get container status \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": rpc error: code = NotFound desc = could not find container \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": container with ID starting with 08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.485877 4993 scope.go:117] "RemoveContainer" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486121 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} err="failed to get container status \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": rpc error: code = NotFound desc = could not find container \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": container with ID starting with 46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486151 4993 scope.go:117] "RemoveContainer" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486394 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} err="failed to get container status \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": rpc error: code = NotFound desc = could not find container \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": container with ID starting with 00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486414 4993 scope.go:117] "RemoveContainer" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486687 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} err="failed to get container status \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": rpc error: code = NotFound desc = could not find container \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": container with ID starting with f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486707 4993 scope.go:117] "RemoveContainer" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486954 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} err="failed to get container status \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": rpc error: code = NotFound desc = could not find container \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": container with ID starting with 16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.486978 4993 scope.go:117] "RemoveContainer" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.487342 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} err="failed to get container status \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": rpc error: code = NotFound desc = could not find container \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": container with ID starting with 19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.487372 4993 scope.go:117] "RemoveContainer" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.487767 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} err="failed to get container status \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": rpc error: code = NotFound desc = could not find container \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": container with ID starting with 119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.487788 4993 scope.go:117] "RemoveContainer" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.488057 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} err="failed to get container status \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": rpc error: code = NotFound desc = could not find container \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": container with ID starting with 771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.488084 4993 scope.go:117] "RemoveContainer" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.488355 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} err="failed to get container status \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": rpc error: code = NotFound desc = could not find container \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": container with ID starting with 0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.488381 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490326 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} err="failed to get container status \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490356 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490667 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} err="failed to get container status \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": rpc error: code = NotFound desc = could not find container \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": container with ID starting with 08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490696 4993 scope.go:117] "RemoveContainer" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490963 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} err="failed to get container status \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": rpc error: code = NotFound desc = could not find container \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": container with ID starting with 46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.490983 4993 scope.go:117] "RemoveContainer" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.491258 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} err="failed to get container status \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": rpc error: code = NotFound desc = could not find container \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": container with ID starting with 00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.491280 4993 scope.go:117] "RemoveContainer" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.491659 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} err="failed to get container status \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": rpc error: code = NotFound desc = could not find container \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": container with ID starting with f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.491677 4993 scope.go:117] "RemoveContainer" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.491995 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} err="failed to get container status \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": rpc error: code = NotFound desc = could not find container \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": container with ID starting with 16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492011 4993 scope.go:117] "RemoveContainer" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492335 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} err="failed to get container status \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": rpc error: code = NotFound desc = could not find container \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": container with ID starting with 19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492352 4993 scope.go:117] "RemoveContainer" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492618 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} err="failed to get container status \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": rpc error: code = NotFound desc = could not find container \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": container with ID starting with 119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492633 4993 scope.go:117] "RemoveContainer" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492906 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} err="failed to get container status \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": rpc error: code = NotFound desc = could not find container \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": container with ID starting with 771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.492930 4993 scope.go:117] "RemoveContainer" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493207 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} err="failed to get container status \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": rpc error: code = NotFound desc = could not find container \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": container with ID starting with 0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493228 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493498 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} err="failed to get container status \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493521 4993 scope.go:117] "RemoveContainer" containerID="08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493792 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef"} err="failed to get container status \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": rpc error: code = NotFound desc = could not find container \"08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef\": container with ID starting with 08636b97c60994965e9ba907a275ed4adb89d24b38117e5bbfc98858ad1e08ef not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.493811 4993 scope.go:117] "RemoveContainer" containerID="46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494042 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828"} err="failed to get container status \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": rpc error: code = NotFound desc = could not find container \"46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828\": container with ID starting with 46c94578648edd768f81ca5826bee86f281b9730b0217b2b00eaa100ea816828 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494056 4993 scope.go:117] "RemoveContainer" containerID="00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494308 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc"} err="failed to get container status \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": rpc error: code = NotFound desc = could not find container \"00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc\": container with ID starting with 00272f7e037918654683d188be5f777359c5a83b8bb1388478da653d00175cfc not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494332 4993 scope.go:117] "RemoveContainer" containerID="f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494624 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa"} err="failed to get container status \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": rpc error: code = NotFound desc = could not find container \"f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa\": container with ID starting with f9061055d073d69a32d61a4ed0dd5dd9b99021a0fc3ce7078f520c9b494c0daa not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.494640 4993 scope.go:117] "RemoveContainer" containerID="16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.497356 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee"} err="failed to get container status \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": rpc error: code = NotFound desc = could not find container \"16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee\": container with ID starting with 16762d5d9b9e2b88f6c868fef67eca871ee6ddf6a07ec04354791c86441b10ee not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.497382 4993 scope.go:117] "RemoveContainer" containerID="19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.498305 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e"} err="failed to get container status \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": rpc error: code = NotFound desc = could not find container \"19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e\": container with ID starting with 19bf8c21f2d13e11b9370480fe05754366c78aecfc0013c9af4a308cec99251e not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.498373 4993 scope.go:117] "RemoveContainer" containerID="119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.498688 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18"} err="failed to get container status \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": rpc error: code = NotFound desc = could not find container \"119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18\": container with ID starting with 119b738636c1ce11a45aac41169cef96101b2a923c1e6a9ca2fb30df8e705e18 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.498712 4993 scope.go:117] "RemoveContainer" containerID="771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.499282 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7"} err="failed to get container status \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": rpc error: code = NotFound desc = could not find container \"771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7\": container with ID starting with 771c5f59b0d2c7b3c4c83749867d32d265fd008a06490b903fc913d11d1f60f7 not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.499333 4993 scope.go:117] "RemoveContainer" containerID="0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.499599 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c"} err="failed to get container status \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": rpc error: code = NotFound desc = could not find container \"0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c\": container with ID starting with 0430381157b672ecb131ad8546319e3ccba01ff3029e21dbd909ef1b50b3843c not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.499624 4993 scope.go:117] "RemoveContainer" containerID="b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.500822 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f"} err="failed to get container status \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": rpc error: code = NotFound desc = could not find container \"b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f\": container with ID starting with b00ad79b3ed6dee565023626331f875ef976710f2d14c7c80890dcfbd7fc059f not found: ID does not exist" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510043 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-netns\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510103 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-var-lib-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510133 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-config\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510224 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovn-node-metrics-cert\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510543 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510650 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46j6h\" (UniqueName: \"kubernetes.io/projected/9cca3ae7-38ba-4f55-8abd-cf2323a15195-kube-api-access-46j6h\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510796 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.510971 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-env-overrides\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511012 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-script-lib\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511067 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-systemd-units\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511091 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-node-log\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511111 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511132 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-ovn\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511152 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-log-socket\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511217 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-slash\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511240 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-systemd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511287 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-etc-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511310 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-netd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511389 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-kubelet\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511414 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-bin\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511615 4993 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511636 4993 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511651 4993 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-log-socket\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511663 4993 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-node-log\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511674 4993 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511685 4993 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511696 4993 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511708 4993 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511721 4993 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-slash\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511732 4993 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511743 4993 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511759 4993 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511770 4993 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511781 4993 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511793 4993 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c6387d35-c6e4-42d4-9f89-6f2940878f8a-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511804 4993 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511815 4993 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511826 4993 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511837 4993 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c6387d35-c6e4-42d4-9f89-6f2940878f8a-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.511847 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t8gs\" (UniqueName: \"kubernetes.io/projected/c6387d35-c6e4-42d4-9f89-6f2940878f8a-kube-api-access-5t8gs\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622652 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-netns\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622722 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-var-lib-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622769 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-config\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622805 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovn-node-metrics-cert\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622851 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622890 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46j6h\" (UniqueName: \"kubernetes.io/projected/9cca3ae7-38ba-4f55-8abd-cf2323a15195-kube-api-access-46j6h\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.622942 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623005 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-env-overrides\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623045 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-script-lib\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623109 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-systemd-units\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623147 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-node-log\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623207 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623240 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-log-socket\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623283 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-slash\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623321 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-systemd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623361 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-ovn\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623408 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-etc-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623442 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-netd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623520 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-kubelet\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623561 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-bin\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623794 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-bin\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623866 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-netns\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.623922 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-var-lib-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.624956 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-config\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625160 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-run-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625206 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-ovn\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625213 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-log-socket\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625213 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-slash\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625247 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-cni-netd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625244 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzr6r"] Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625321 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-systemd-units\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625322 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625360 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-run-systemd\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625356 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625387 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-etc-openvswitch\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625392 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-node-log\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.625410 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cca3ae7-38ba-4f55-8abd-cf2323a15195-host-kubelet\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.626017 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-env-overrides\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.626007 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovnkube-script-lib\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.629270 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzr6r"] Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.630082 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cca3ae7-38ba-4f55-8abd-cf2323a15195-ovn-node-metrics-cert\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.643674 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46j6h\" (UniqueName: \"kubernetes.io/projected/9cca3ae7-38ba-4f55-8abd-cf2323a15195-kube-api-access-46j6h\") pod \"ovnkube-node-mx5jx\" (UID: \"9cca3ae7-38ba-4f55-8abd-cf2323a15195\") " pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: I0129 12:19:05.669695 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:05 crc kubenswrapper[4993]: W0129 12:19:05.690197 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cca3ae7_38ba_4f55_8abd_cf2323a15195.slice/crio-0f41e4d11b37f0edf64b0dca80ef8670f4dc86940ec3ba83a373110d0c7bd986 WatchSource:0}: Error finding container 0f41e4d11b37f0edf64b0dca80ef8670f4dc86940ec3ba83a373110d0c7bd986: Status 404 returned error can't find the container with id 0f41e4d11b37f0edf64b0dca80ef8670f4dc86940ec3ba83a373110d0c7bd986 Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.300381 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/2.log" Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.301702 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/1.log" Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.301786 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b9gn8" event={"ID":"d22ea793-b0f9-4b2c-8ea7-ed61728d900f","Type":"ContainerStarted","Data":"e5e1a86f15f3dd733898bdd628a40b1f87f4e59839313ed2d1f7d7ea1b2b35de"} Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.304364 4993 generic.go:334] "Generic (PLEG): container finished" podID="9cca3ae7-38ba-4f55-8abd-cf2323a15195" containerID="e8f2651cc79d78817808c4d96b70b966978639675cdeeb4de5eb007da7557bec" exitCode=0 Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.304393 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerDied","Data":"e8f2651cc79d78817808c4d96b70b966978639675cdeeb4de5eb007da7557bec"} Jan 29 12:19:06 crc kubenswrapper[4993]: I0129 12:19:06.304410 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"0f41e4d11b37f0edf64b0dca80ef8670f4dc86940ec3ba83a373110d0c7bd986"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.196271 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6387d35-c6e4-42d4-9f89-6f2940878f8a" path="/var/lib/kubelet/pods/c6387d35-c6e4-42d4-9f89-6f2940878f8a/volumes" Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312710 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"a6b52390b3d66787f397caac9b392f0c69fc7f891bfbc804f508bf4c5b6c26af"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312760 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"2def150863098ce32de9b76f25183195dcbb3cc8c2ee29a31f9bf651e626d7ad"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312774 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"b9cb7f50bf0e31a4735b1d8b760299b7dd15d2a482900b4cfb9862dbeb8984de"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312789 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"0940e49b6908bec470743d13f86968614105f0679812134f95c764be08a21208"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312805 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"5ffd8aa1380d75e64f8656165e6f7d7a58257bd0e2f8359a5cb45a94fc0c3493"} Jan 29 12:19:07 crc kubenswrapper[4993]: I0129 12:19:07.312816 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"1cb96c0abe48c4160030745b27d8313aebb48524f98bfad8b573ab5ad673ecb2"} Jan 29 12:19:09 crc kubenswrapper[4993]: I0129 12:19:09.326771 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"d04dbb230bfcc19d92c24a3e99ec689544215ecbb4ebc9eb8b9824184f0bb958"} Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.345021 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" event={"ID":"9cca3ae7-38ba-4f55-8abd-cf2323a15195","Type":"ContainerStarted","Data":"6ef12c7d377f2e5fbdde44f41e5c2da5ff3608eb7cabfd25be19e33db7847892"} Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.345932 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.345954 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.345967 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.379669 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.402643 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:12 crc kubenswrapper[4993]: I0129 12:19:12.416298 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" podStartSLOduration=7.416280248 podStartE2EDuration="7.416280248s" podCreationTimestamp="2026-01-29 12:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:19:12.383561208 +0000 UTC m=+836.398691354" watchObservedRunningTime="2026-01-29 12:19:12.416280248 +0000 UTC m=+836.431410374" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.586526 4993 scope.go:117] "RemoveContainer" containerID="cefd82ce2e4ec8129fe57b15712637983e265830f3f36327ea043af657feef5b" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.950061 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-tn5rp"] Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.951405 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.953554 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.954095 4993 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-jz87c" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.954273 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.956074 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.959225 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tn5rp"] Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.975152 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.975233 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:17 crc kubenswrapper[4993]: I0129 12:19:17.975315 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7nnk\" (UniqueName: \"kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.077406 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7nnk\" (UniqueName: \"kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.077566 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.077607 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.078080 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.078621 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.099739 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7nnk\" (UniqueName: \"kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk\") pod \"crc-storage-crc-tn5rp\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.271648 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.377694 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b9gn8_d22ea793-b0f9-4b2c-8ea7-ed61728d900f/kube-multus/2.log" Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.456016 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-tn5rp"] Jan 29 12:19:18 crc kubenswrapper[4993]: W0129 12:19:18.462291 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb4e2365_e8ef_4777_ae16_c39c2076cb41.slice/crio-d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261 WatchSource:0}: Error finding container d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261: Status 404 returned error can't find the container with id d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261 Jan 29 12:19:18 crc kubenswrapper[4993]: I0129 12:19:18.464335 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:19:19 crc kubenswrapper[4993]: I0129 12:19:19.383078 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tn5rp" event={"ID":"cb4e2365-e8ef-4777-ae16-c39c2076cb41","Type":"ContainerStarted","Data":"d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261"} Jan 29 12:19:20 crc kubenswrapper[4993]: I0129 12:19:20.390623 4993 generic.go:334] "Generic (PLEG): container finished" podID="cb4e2365-e8ef-4777-ae16-c39c2076cb41" containerID="dca3393aa9e41776d60f7f3c2a33064b61f162befa71dec1b9a0dd89dba2e15c" exitCode=0 Jan 29 12:19:20 crc kubenswrapper[4993]: I0129 12:19:20.390753 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tn5rp" event={"ID":"cb4e2365-e8ef-4777-ae16-c39c2076cb41","Type":"ContainerDied","Data":"dca3393aa9e41776d60f7f3c2a33064b61f162befa71dec1b9a0dd89dba2e15c"} Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.611519 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.618432 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt\") pod \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.618532 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7nnk\" (UniqueName: \"kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk\") pod \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.618603 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "cb4e2365-e8ef-4777-ae16-c39c2076cb41" (UID: "cb4e2365-e8ef-4777-ae16-c39c2076cb41"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.618625 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage\") pod \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\" (UID: \"cb4e2365-e8ef-4777-ae16-c39c2076cb41\") " Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.619408 4993 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb4e2365-e8ef-4777-ae16-c39c2076cb41-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.627665 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk" (OuterVolumeSpecName: "kube-api-access-w7nnk") pod "cb4e2365-e8ef-4777-ae16-c39c2076cb41" (UID: "cb4e2365-e8ef-4777-ae16-c39c2076cb41"). InnerVolumeSpecName "kube-api-access-w7nnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.637838 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "cb4e2365-e8ef-4777-ae16-c39c2076cb41" (UID: "cb4e2365-e8ef-4777-ae16-c39c2076cb41"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.720437 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7nnk\" (UniqueName: \"kubernetes.io/projected/cb4e2365-e8ef-4777-ae16-c39c2076cb41-kube-api-access-w7nnk\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:21 crc kubenswrapper[4993]: I0129 12:19:21.720468 4993 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb4e2365-e8ef-4777-ae16-c39c2076cb41-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:22 crc kubenswrapper[4993]: I0129 12:19:22.402150 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-tn5rp" event={"ID":"cb4e2365-e8ef-4777-ae16-c39c2076cb41","Type":"ContainerDied","Data":"d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261"} Jan 29 12:19:22 crc kubenswrapper[4993]: I0129 12:19:22.402707 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d53247bd68a04364f8f82fb37b6a28877b5b15018dbbfbd0dba03a13d8d1f261" Jan 29 12:19:22 crc kubenswrapper[4993]: I0129 12:19:22.402403 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-tn5rp" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.262545 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26"] Jan 29 12:19:28 crc kubenswrapper[4993]: E0129 12:19:28.264121 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4e2365-e8ef-4777-ae16-c39c2076cb41" containerName="storage" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.264266 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4e2365-e8ef-4777-ae16-c39c2076cb41" containerName="storage" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.264513 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4e2365-e8ef-4777-ae16-c39c2076cb41" containerName="storage" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.265538 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.274050 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.276414 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26"] Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.297925 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.297976 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.298076 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6wf9\" (UniqueName: \"kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.399520 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6wf9\" (UniqueName: \"kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.399595 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.399627 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.400180 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.400268 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.418108 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6wf9\" (UniqueName: \"kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.596843 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:28 crc kubenswrapper[4993]: I0129 12:19:28.990009 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26"] Jan 29 12:19:28 crc kubenswrapper[4993]: W0129 12:19:28.998278 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod580e9990_2a5c_4761_b9a6_1111d922725c.slice/crio-fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9 WatchSource:0}: Error finding container fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9: Status 404 returned error can't find the container with id fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9 Jan 29 12:19:29 crc kubenswrapper[4993]: I0129 12:19:29.436272 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerStarted","Data":"6243d1725dffdae56a6404efbebbf9e19a6a08f90b8db91c7b42e8a853d21b31"} Jan 29 12:19:29 crc kubenswrapper[4993]: I0129 12:19:29.436530 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerStarted","Data":"fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9"} Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.369159 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.370872 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.382302 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.421842 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.421889 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn6df\" (UniqueName: \"kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.421932 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.441351 4993 generic.go:334] "Generic (PLEG): container finished" podID="580e9990-2a5c-4761-b9a6-1111d922725c" containerID="6243d1725dffdae56a6404efbebbf9e19a6a08f90b8db91c7b42e8a853d21b31" exitCode=0 Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.441393 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerDied","Data":"6243d1725dffdae56a6404efbebbf9e19a6a08f90b8db91c7b42e8a853d21b31"} Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.522856 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.522912 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn6df\" (UniqueName: \"kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.523063 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.523542 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.523550 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.542606 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn6df\" (UniqueName: \"kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df\") pod \"redhat-operators-wfbbz\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.688882 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:30 crc kubenswrapper[4993]: I0129 12:19:30.891984 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:30 crc kubenswrapper[4993]: W0129 12:19:30.906957 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f8194ab_e750_42c3_b6ee_dfdd00c0f9d3.slice/crio-86829ec9d7baf20adb0f5992181950b8c3aa1bb1adacf5d8d4ea86d0ca3a7886 WatchSource:0}: Error finding container 86829ec9d7baf20adb0f5992181950b8c3aa1bb1adacf5d8d4ea86d0ca3a7886: Status 404 returned error can't find the container with id 86829ec9d7baf20adb0f5992181950b8c3aa1bb1adacf5d8d4ea86d0ca3a7886 Jan 29 12:19:31 crc kubenswrapper[4993]: I0129 12:19:31.452108 4993 generic.go:334] "Generic (PLEG): container finished" podID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerID="443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b" exitCode=0 Jan 29 12:19:31 crc kubenswrapper[4993]: I0129 12:19:31.452220 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerDied","Data":"443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b"} Jan 29 12:19:31 crc kubenswrapper[4993]: I0129 12:19:31.452259 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerStarted","Data":"86829ec9d7baf20adb0f5992181950b8c3aa1bb1adacf5d8d4ea86d0ca3a7886"} Jan 29 12:19:32 crc kubenswrapper[4993]: I0129 12:19:32.459294 4993 generic.go:334] "Generic (PLEG): container finished" podID="580e9990-2a5c-4761-b9a6-1111d922725c" containerID="b7335592e8afe1a356cace6f6e0bb094dd60781dc2375922d52571549af5594d" exitCode=0 Jan 29 12:19:32 crc kubenswrapper[4993]: I0129 12:19:32.459365 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerDied","Data":"b7335592e8afe1a356cace6f6e0bb094dd60781dc2375922d52571549af5594d"} Jan 29 12:19:33 crc kubenswrapper[4993]: I0129 12:19:33.468298 4993 generic.go:334] "Generic (PLEG): container finished" podID="580e9990-2a5c-4761-b9a6-1111d922725c" containerID="e8948ed3e5f82e9ba43a1a026af44518d678a1877a740d1c7d1b16667b8c377a" exitCode=0 Jan 29 12:19:33 crc kubenswrapper[4993]: I0129 12:19:33.468391 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerDied","Data":"e8948ed3e5f82e9ba43a1a026af44518d678a1877a740d1c7d1b16667b8c377a"} Jan 29 12:19:33 crc kubenswrapper[4993]: I0129 12:19:33.471048 4993 generic.go:334] "Generic (PLEG): container finished" podID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerID="68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e" exitCode=0 Jan 29 12:19:33 crc kubenswrapper[4993]: I0129 12:19:33.471106 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerDied","Data":"68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e"} Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.478596 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerStarted","Data":"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b"} Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.497634 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wfbbz" podStartSLOduration=2.027788306 podStartE2EDuration="4.497614845s" podCreationTimestamp="2026-01-29 12:19:30 +0000 UTC" firstStartedPulling="2026-01-29 12:19:31.453441572 +0000 UTC m=+855.468571698" lastFinishedPulling="2026-01-29 12:19:33.923268111 +0000 UTC m=+857.938398237" observedRunningTime="2026-01-29 12:19:34.495558101 +0000 UTC m=+858.510688237" watchObservedRunningTime="2026-01-29 12:19:34.497614845 +0000 UTC m=+858.512744971" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.712896 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.873629 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle\") pod \"580e9990-2a5c-4761-b9a6-1111d922725c\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.873743 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util\") pod \"580e9990-2a5c-4761-b9a6-1111d922725c\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.873869 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6wf9\" (UniqueName: \"kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9\") pod \"580e9990-2a5c-4761-b9a6-1111d922725c\" (UID: \"580e9990-2a5c-4761-b9a6-1111d922725c\") " Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.874367 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle" (OuterVolumeSpecName: "bundle") pod "580e9990-2a5c-4761-b9a6-1111d922725c" (UID: "580e9990-2a5c-4761-b9a6-1111d922725c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.880345 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9" (OuterVolumeSpecName: "kube-api-access-g6wf9") pod "580e9990-2a5c-4761-b9a6-1111d922725c" (UID: "580e9990-2a5c-4761-b9a6-1111d922725c"). InnerVolumeSpecName "kube-api-access-g6wf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.884555 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util" (OuterVolumeSpecName: "util") pod "580e9990-2a5c-4761-b9a6-1111d922725c" (UID: "580e9990-2a5c-4761-b9a6-1111d922725c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.975058 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6wf9\" (UniqueName: \"kubernetes.io/projected/580e9990-2a5c-4761-b9a6-1111d922725c-kube-api-access-g6wf9\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.975092 4993 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:34 crc kubenswrapper[4993]: I0129 12:19:34.975101 4993 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580e9990-2a5c-4761-b9a6-1111d922725c-util\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:35 crc kubenswrapper[4993]: I0129 12:19:35.486675 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" event={"ID":"580e9990-2a5c-4761-b9a6-1111d922725c","Type":"ContainerDied","Data":"fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9"} Jan 29 12:19:35 crc kubenswrapper[4993]: I0129 12:19:35.486704 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26" Jan 29 12:19:35 crc kubenswrapper[4993]: I0129 12:19:35.486723 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe4a7bfaec1aba9b0de72a3fb0f6a020d46ad25be552a93d1a270a1dcffb8cc9" Jan 29 12:19:35 crc kubenswrapper[4993]: I0129 12:19:35.691538 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mx5jx" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.532612 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-s54j9"] Jan 29 12:19:39 crc kubenswrapper[4993]: E0129 12:19:39.533845 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="util" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.533865 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="util" Jan 29 12:19:39 crc kubenswrapper[4993]: E0129 12:19:39.533879 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="pull" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.533886 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="pull" Jan 29 12:19:39 crc kubenswrapper[4993]: E0129 12:19:39.533904 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="extract" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.533913 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="extract" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.534025 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="580e9990-2a5c-4761-b9a6-1111d922725c" containerName="extract" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.534781 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.537919 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.538471 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.539718 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-k68nr" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.544491 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-s54j9"] Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.734976 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24l2z\" (UniqueName: \"kubernetes.io/projected/5719c06f-4107-47e6-af46-cc7fcc25d3d8-kube-api-access-24l2z\") pod \"nmstate-operator-646758c888-s54j9\" (UID: \"5719c06f-4107-47e6-af46-cc7fcc25d3d8\") " pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.836004 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24l2z\" (UniqueName: \"kubernetes.io/projected/5719c06f-4107-47e6-af46-cc7fcc25d3d8-kube-api-access-24l2z\") pod \"nmstate-operator-646758c888-s54j9\" (UID: \"5719c06f-4107-47e6-af46-cc7fcc25d3d8\") " pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.864073 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24l2z\" (UniqueName: \"kubernetes.io/projected/5719c06f-4107-47e6-af46-cc7fcc25d3d8-kube-api-access-24l2z\") pod \"nmstate-operator-646758c888-s54j9\" (UID: \"5719c06f-4107-47e6-af46-cc7fcc25d3d8\") " pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" Jan 29 12:19:39 crc kubenswrapper[4993]: I0129 12:19:39.927070 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" Jan 29 12:19:40 crc kubenswrapper[4993]: I0129 12:19:40.349162 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-s54j9"] Jan 29 12:19:40 crc kubenswrapper[4993]: W0129 12:19:40.353767 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5719c06f_4107_47e6_af46_cc7fcc25d3d8.slice/crio-2fc31ed1584d218ff05fc7d2a520c018b291847e51a5cdea272721636cb9efd3 WatchSource:0}: Error finding container 2fc31ed1584d218ff05fc7d2a520c018b291847e51a5cdea272721636cb9efd3: Status 404 returned error can't find the container with id 2fc31ed1584d218ff05fc7d2a520c018b291847e51a5cdea272721636cb9efd3 Jan 29 12:19:40 crc kubenswrapper[4993]: I0129 12:19:40.512362 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" event={"ID":"5719c06f-4107-47e6-af46-cc7fcc25d3d8","Type":"ContainerStarted","Data":"2fc31ed1584d218ff05fc7d2a520c018b291847e51a5cdea272721636cb9efd3"} Jan 29 12:19:40 crc kubenswrapper[4993]: I0129 12:19:40.689788 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:40 crc kubenswrapper[4993]: I0129 12:19:40.690510 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:40 crc kubenswrapper[4993]: I0129 12:19:40.763610 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:41 crc kubenswrapper[4993]: I0129 12:19:41.573679 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:43 crc kubenswrapper[4993]: I0129 12:19:43.364001 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:43 crc kubenswrapper[4993]: I0129 12:19:43.530436 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wfbbz" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="registry-server" containerID="cri-o://a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b" gracePeriod=2 Jan 29 12:19:43 crc kubenswrapper[4993]: I0129 12:19:43.930411 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.091607 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn6df\" (UniqueName: \"kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df\") pod \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.092213 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities\") pod \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.092533 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content\") pod \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\" (UID: \"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3\") " Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.093126 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities" (OuterVolumeSpecName: "utilities") pod "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" (UID: "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.097686 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df" (OuterVolumeSpecName: "kube-api-access-rn6df") pod "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" (UID: "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3"). InnerVolumeSpecName "kube-api-access-rn6df". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.193742 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn6df\" (UniqueName: \"kubernetes.io/projected/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-kube-api-access-rn6df\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.193770 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.234762 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" (UID: "3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.296744 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.541616 4993 generic.go:334] "Generic (PLEG): container finished" podID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerID="a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b" exitCode=0 Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.541756 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wfbbz" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.541755 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerDied","Data":"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b"} Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.543902 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wfbbz" event={"ID":"3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3","Type":"ContainerDied","Data":"86829ec9d7baf20adb0f5992181950b8c3aa1bb1adacf5d8d4ea86d0ca3a7886"} Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.543942 4993 scope.go:117] "RemoveContainer" containerID="a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.547640 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" event={"ID":"5719c06f-4107-47e6-af46-cc7fcc25d3d8","Type":"ContainerStarted","Data":"40ae2036ad51a4e02e53d174f03d8f38d2c00d93ab4a789570000f808f9feff1"} Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.573934 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-s54j9" podStartSLOduration=2.470303554 podStartE2EDuration="5.573907386s" podCreationTimestamp="2026-01-29 12:19:39 +0000 UTC" firstStartedPulling="2026-01-29 12:19:40.356694798 +0000 UTC m=+864.371824914" lastFinishedPulling="2026-01-29 12:19:43.46029853 +0000 UTC m=+867.475428746" observedRunningTime="2026-01-29 12:19:44.570979447 +0000 UTC m=+868.586109573" watchObservedRunningTime="2026-01-29 12:19:44.573907386 +0000 UTC m=+868.589037512" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.575424 4993 scope.go:117] "RemoveContainer" containerID="68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.593949 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.609464 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wfbbz"] Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.630043 4993 scope.go:117] "RemoveContainer" containerID="443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.660523 4993 scope.go:117] "RemoveContainer" containerID="a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b" Jan 29 12:19:44 crc kubenswrapper[4993]: E0129 12:19:44.664411 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b\": container with ID starting with a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b not found: ID does not exist" containerID="a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.664475 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b"} err="failed to get container status \"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b\": rpc error: code = NotFound desc = could not find container \"a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b\": container with ID starting with a916b8f245f163f670957da525fda8c31ada73a3827e07ea9ca2a70a54e51f9b not found: ID does not exist" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.664512 4993 scope.go:117] "RemoveContainer" containerID="68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e" Jan 29 12:19:44 crc kubenswrapper[4993]: E0129 12:19:44.668450 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e\": container with ID starting with 68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e not found: ID does not exist" containerID="68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.668526 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e"} err="failed to get container status \"68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e\": rpc error: code = NotFound desc = could not find container \"68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e\": container with ID starting with 68fa18307d99188dc0a6fad499bdedce31c5329e63e8910cab8ccc3852d5ae8e not found: ID does not exist" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.668569 4993 scope.go:117] "RemoveContainer" containerID="443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b" Jan 29 12:19:44 crc kubenswrapper[4993]: E0129 12:19:44.673455 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b\": container with ID starting with 443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b not found: ID does not exist" containerID="443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b" Jan 29 12:19:44 crc kubenswrapper[4993]: I0129 12:19:44.673533 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b"} err="failed to get container status \"443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b\": rpc error: code = NotFound desc = could not find container \"443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b\": container with ID starting with 443b10df7a70c5d5799f1a16195eada81489abb42463cc7034bf63571686401b not found: ID does not exist" Jan 29 12:19:45 crc kubenswrapper[4993]: I0129 12:19:45.200563 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" path="/var/lib/kubelet/pods/3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3/volumes" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.367996 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvw7j"] Jan 29 12:19:48 crc kubenswrapper[4993]: E0129 12:19:48.369217 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="extract-content" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.369242 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="extract-content" Jan 29 12:19:48 crc kubenswrapper[4993]: E0129 12:19:48.369267 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="registry-server" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.369281 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="registry-server" Jan 29 12:19:48 crc kubenswrapper[4993]: E0129 12:19:48.369310 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="extract-utilities" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.369324 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="extract-utilities" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.369657 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f8194ab-e750-42c3-b6ee-dfdd00c0f9d3" containerName="registry-server" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.378810 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.394817 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-7zqxj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.396368 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.397273 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.399581 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.401948 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvw7j"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.411974 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-tzsvj"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.413980 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.416134 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.455862 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2v22\" (UniqueName: \"kubernetes.io/projected/b407b3c8-24b5-400b-967c-f8145f41bc4b-kube-api-access-b2v22\") pod \"nmstate-metrics-54757c584b-vvw7j\" (UID: \"b407b3c8-24b5-400b-967c-f8145f41bc4b\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.543826 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.544757 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.546789 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.548986 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.549146 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-r92kr" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.554599 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557095 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-nmstate-lock\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557397 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/588bc834-7403-44b7-8152-7a18d3da5e92-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557518 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/dd99b0bb-57cb-4034-bf70-c27b85d936d2-kube-api-access-fxjbr\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557597 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2v22\" (UniqueName: \"kubernetes.io/projected/b407b3c8-24b5-400b-967c-f8145f41bc4b-kube-api-access-b2v22\") pod \"nmstate-metrics-54757c584b-vvw7j\" (UID: \"b407b3c8-24b5-400b-967c-f8145f41bc4b\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557631 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-ovs-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557684 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-dbus-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.557852 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zps62\" (UniqueName: \"kubernetes.io/projected/588bc834-7403-44b7-8152-7a18d3da5e92-kube-api-access-zps62\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.582716 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2v22\" (UniqueName: \"kubernetes.io/projected/b407b3c8-24b5-400b-967c-f8145f41bc4b-kube-api-access-b2v22\") pod \"nmstate-metrics-54757c584b-vvw7j\" (UID: \"b407b3c8-24b5-400b-967c-f8145f41bc4b\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659037 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-nmstate-lock\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659126 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/588bc834-7403-44b7-8152-7a18d3da5e92-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659165 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/dd99b0bb-57cb-4034-bf70-c27b85d936d2-kube-api-access-fxjbr\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659222 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-ovs-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659251 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-dbus-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659283 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp9b2\" (UniqueName: \"kubernetes.io/projected/24b06982-4e48-47f3-84ed-8fd0f34b1f43-kube-api-access-vp9b2\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659334 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/24b06982-4e48-47f3-84ed-8fd0f34b1f43-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659373 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/24b06982-4e48-47f3-84ed-8fd0f34b1f43-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659407 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zps62\" (UniqueName: \"kubernetes.io/projected/588bc834-7403-44b7-8152-7a18d3da5e92-kube-api-access-zps62\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659691 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-ovs-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659876 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-nmstate-lock\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.659891 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/dd99b0bb-57cb-4034-bf70-c27b85d936d2-dbus-socket\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.667313 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/588bc834-7403-44b7-8152-7a18d3da5e92-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.679809 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zps62\" (UniqueName: \"kubernetes.io/projected/588bc834-7403-44b7-8152-7a18d3da5e92-kube-api-access-zps62\") pod \"nmstate-webhook-8474b5b9d8-2hphw\" (UID: \"588bc834-7403-44b7-8152-7a18d3da5e92\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.694365 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxjbr\" (UniqueName: \"kubernetes.io/projected/dd99b0bb-57cb-4034-bf70-c27b85d936d2-kube-api-access-fxjbr\") pod \"nmstate-handler-tzsvj\" (UID: \"dd99b0bb-57cb-4034-bf70-c27b85d936d2\") " pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.726483 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.735587 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.741880 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-68cc7b54cf-dl89k"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.742540 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.746464 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.760094 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp9b2\" (UniqueName: \"kubernetes.io/projected/24b06982-4e48-47f3-84ed-8fd0f34b1f43-kube-api-access-vp9b2\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.760157 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/24b06982-4e48-47f3-84ed-8fd0f34b1f43-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.760220 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/24b06982-4e48-47f3-84ed-8fd0f34b1f43-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.761260 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/24b06982-4e48-47f3-84ed-8fd0f34b1f43-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.766163 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68cc7b54cf-dl89k"] Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.767930 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/24b06982-4e48-47f3-84ed-8fd0f34b1f43-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.785160 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp9b2\" (UniqueName: \"kubernetes.io/projected/24b06982-4e48-47f3-84ed-8fd0f34b1f43-kube-api-access-vp9b2\") pod \"nmstate-console-plugin-7754f76f8b-vn2fg\" (UID: \"24b06982-4e48-47f3-84ed-8fd0f34b1f43\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.860957 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-console-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861057 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-trusted-ca-bundle\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861099 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861240 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgmm\" (UniqueName: \"kubernetes.io/projected/7c04c146-f887-4cec-a0f0-483351888bb1-kube-api-access-6tgmm\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861286 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-oauth-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861303 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-oauth-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.861352 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-service-ca\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.873523 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962240 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-trusted-ca-bundle\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962609 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962642 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgmm\" (UniqueName: \"kubernetes.io/projected/7c04c146-f887-4cec-a0f0-483351888bb1-kube-api-access-6tgmm\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962657 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-oauth-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962673 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-oauth-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962693 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-service-ca\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.962725 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-console-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.964105 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-service-ca\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.964168 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-console-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.964249 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-oauth-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.964912 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c04c146-f887-4cec-a0f0-483351888bb1-trusted-ca-bundle\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.968080 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-serving-cert\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.968104 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c04c146-f887-4cec-a0f0-483351888bb1-console-oauth-config\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:48 crc kubenswrapper[4993]: I0129 12:19:48.984778 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgmm\" (UniqueName: \"kubernetes.io/projected/7c04c146-f887-4cec-a0f0-483351888bb1-kube-api-access-6tgmm\") pod \"console-68cc7b54cf-dl89k\" (UID: \"7c04c146-f887-4cec-a0f0-483351888bb1\") " pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.077954 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg"] Jan 29 12:19:49 crc kubenswrapper[4993]: W0129 12:19:49.082023 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24b06982_4e48_47f3_84ed_8fd0f34b1f43.slice/crio-5a2301329e1f1afbc7dddd40faa953b0d67b3fcfae013ad54e55cafedfaaedef WatchSource:0}: Error finding container 5a2301329e1f1afbc7dddd40faa953b0d67b3fcfae013ad54e55cafedfaaedef: Status 404 returned error can't find the container with id 5a2301329e1f1afbc7dddd40faa953b0d67b3fcfae013ad54e55cafedfaaedef Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.089726 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.142051 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw"] Jan 29 12:19:49 crc kubenswrapper[4993]: W0129 12:19:49.150962 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod588bc834_7403_44b7_8152_7a18d3da5e92.slice/crio-7a55641514b0705592f2badb8213750f954a234b4162daac1c114207d51072c2 WatchSource:0}: Error finding container 7a55641514b0705592f2badb8213750f954a234b4162daac1c114207d51072c2: Status 404 returned error can't find the container with id 7a55641514b0705592f2badb8213750f954a234b4162daac1c114207d51072c2 Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.200203 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvw7j"] Jan 29 12:19:49 crc kubenswrapper[4993]: W0129 12:19:49.207435 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb407b3c8_24b5_400b_967c_f8145f41bc4b.slice/crio-cde1da73ab34105205a1ff6d55e9fda2de3983795be3246a3481aaf75e0cb81c WatchSource:0}: Error finding container cde1da73ab34105205a1ff6d55e9fda2de3983795be3246a3481aaf75e0cb81c: Status 404 returned error can't find the container with id cde1da73ab34105205a1ff6d55e9fda2de3983795be3246a3481aaf75e0cb81c Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.283279 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68cc7b54cf-dl89k"] Jan 29 12:19:49 crc kubenswrapper[4993]: W0129 12:19:49.292726 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c04c146_f887_4cec_a0f0_483351888bb1.slice/crio-2195cad349c1daca93a68c4275a9af412f5bccc8a92dde604b56e1b0758be564 WatchSource:0}: Error finding container 2195cad349c1daca93a68c4275a9af412f5bccc8a92dde604b56e1b0758be564: Status 404 returned error can't find the container with id 2195cad349c1daca93a68c4275a9af412f5bccc8a92dde604b56e1b0758be564 Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.582450 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" event={"ID":"b407b3c8-24b5-400b-967c-f8145f41bc4b","Type":"ContainerStarted","Data":"cde1da73ab34105205a1ff6d55e9fda2de3983795be3246a3481aaf75e0cb81c"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.584096 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" event={"ID":"588bc834-7403-44b7-8152-7a18d3da5e92","Type":"ContainerStarted","Data":"7a55641514b0705592f2badb8213750f954a234b4162daac1c114207d51072c2"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.585045 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" event={"ID":"24b06982-4e48-47f3-84ed-8fd0f34b1f43","Type":"ContainerStarted","Data":"5a2301329e1f1afbc7dddd40faa953b0d67b3fcfae013ad54e55cafedfaaedef"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.586131 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tzsvj" event={"ID":"dd99b0bb-57cb-4034-bf70-c27b85d936d2","Type":"ContainerStarted","Data":"f3cb11172c4bf1362b84dc68ec32a019510b84f04f0dffbd5821e06a53530dfa"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.587246 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68cc7b54cf-dl89k" event={"ID":"7c04c146-f887-4cec-a0f0-483351888bb1","Type":"ContainerStarted","Data":"de3e141dcbd79a88f8929b8b87ef76e3bdbbd3c71a61852f172e669b3a385a0f"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.587276 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68cc7b54cf-dl89k" event={"ID":"7c04c146-f887-4cec-a0f0-483351888bb1","Type":"ContainerStarted","Data":"2195cad349c1daca93a68c4275a9af412f5bccc8a92dde604b56e1b0758be564"} Jan 29 12:19:49 crc kubenswrapper[4993]: I0129 12:19:49.623644 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-68cc7b54cf-dl89k" podStartSLOduration=1.623624728 podStartE2EDuration="1.623624728s" podCreationTimestamp="2026-01-29 12:19:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:19:49.623570247 +0000 UTC m=+873.638700383" watchObservedRunningTime="2026-01-29 12:19:49.623624728 +0000 UTC m=+873.638754854" Jan 29 12:19:51 crc kubenswrapper[4993]: I0129 12:19:51.616736 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" event={"ID":"24b06982-4e48-47f3-84ed-8fd0f34b1f43","Type":"ContainerStarted","Data":"ec349775704ec6760a7cc215c2ba1538f3f9d0f9dd1c849a3884de40e242f717"} Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.624293 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" event={"ID":"588bc834-7403-44b7-8152-7a18d3da5e92","Type":"ContainerStarted","Data":"6406e4449958d3f3fa766eb5862b4d5617c3d9c3add0739494e61f2008ca1779"} Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.624614 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.625859 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tzsvj" event={"ID":"dd99b0bb-57cb-4034-bf70-c27b85d936d2","Type":"ContainerStarted","Data":"c21cc8493b51f9232846cf7d8268771a9c7a2d8ab0c9be378bb71ec0b00e30a7"} Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.625934 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.627674 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" event={"ID":"b407b3c8-24b5-400b-967c-f8145f41bc4b","Type":"ContainerStarted","Data":"dc839713def4ee41b7cb66df26cb604fcae6db670a597eb39e0607ef62e6536c"} Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.655060 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" podStartSLOduration=1.63612118 podStartE2EDuration="4.655032842s" podCreationTimestamp="2026-01-29 12:19:48 +0000 UTC" firstStartedPulling="2026-01-29 12:19:49.177350788 +0000 UTC m=+873.192480914" lastFinishedPulling="2026-01-29 12:19:52.19626245 +0000 UTC m=+876.211392576" observedRunningTime="2026-01-29 12:19:52.650504322 +0000 UTC m=+876.665634468" watchObservedRunningTime="2026-01-29 12:19:52.655032842 +0000 UTC m=+876.670162968" Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.656424 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-vn2fg" podStartSLOduration=2.53692554 podStartE2EDuration="4.656414169s" podCreationTimestamp="2026-01-29 12:19:48 +0000 UTC" firstStartedPulling="2026-01-29 12:19:49.084297885 +0000 UTC m=+873.099428011" lastFinishedPulling="2026-01-29 12:19:51.203786504 +0000 UTC m=+875.218916640" observedRunningTime="2026-01-29 12:19:51.634728547 +0000 UTC m=+875.649858673" watchObservedRunningTime="2026-01-29 12:19:52.656414169 +0000 UTC m=+876.671544295" Jan 29 12:19:52 crc kubenswrapper[4993]: I0129 12:19:52.684858 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-tzsvj" podStartSLOduration=1.310348743 podStartE2EDuration="4.684840535s" podCreationTimestamp="2026-01-29 12:19:48 +0000 UTC" firstStartedPulling="2026-01-29 12:19:48.809613595 +0000 UTC m=+872.824743721" lastFinishedPulling="2026-01-29 12:19:52.184105387 +0000 UTC m=+876.199235513" observedRunningTime="2026-01-29 12:19:52.681079855 +0000 UTC m=+876.696209981" watchObservedRunningTime="2026-01-29 12:19:52.684840535 +0000 UTC m=+876.699970661" Jan 29 12:19:54 crc kubenswrapper[4993]: I0129 12:19:54.640130 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" event={"ID":"b407b3c8-24b5-400b-967c-f8145f41bc4b","Type":"ContainerStarted","Data":"0ee28694a6c7db47016148dd8dbf26e0faa9f9e276f32ed6ddb9c1b25de571a9"} Jan 29 12:19:54 crc kubenswrapper[4993]: I0129 12:19:54.657813 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvw7j" podStartSLOduration=1.6621081009999998 podStartE2EDuration="6.657793138s" podCreationTimestamp="2026-01-29 12:19:48 +0000 UTC" firstStartedPulling="2026-01-29 12:19:49.224378848 +0000 UTC m=+873.239508974" lastFinishedPulling="2026-01-29 12:19:54.220063885 +0000 UTC m=+878.235194011" observedRunningTime="2026-01-29 12:19:54.656564256 +0000 UTC m=+878.671694402" watchObservedRunningTime="2026-01-29 12:19:54.657793138 +0000 UTC m=+878.672923264" Jan 29 12:19:58 crc kubenswrapper[4993]: I0129 12:19:58.771841 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-tzsvj" Jan 29 12:19:59 crc kubenswrapper[4993]: I0129 12:19:59.090657 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:59 crc kubenswrapper[4993]: I0129 12:19:59.090842 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:59 crc kubenswrapper[4993]: I0129 12:19:59.095067 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:59 crc kubenswrapper[4993]: I0129 12:19:59.676070 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-68cc7b54cf-dl89k" Jan 29 12:19:59 crc kubenswrapper[4993]: I0129 12:19:59.734973 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:20:08 crc kubenswrapper[4993]: I0129 12:20:08.744719 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-2hphw" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.553126 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.555414 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.571155 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.596714 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.596879 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.596971 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dmp8\" (UniqueName: \"kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.698762 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.698822 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dmp8\" (UniqueName: \"kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.698906 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.699617 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.699685 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.732242 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dmp8\" (UniqueName: \"kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8\") pod \"redhat-marketplace-sl7zz\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:11 crc kubenswrapper[4993]: I0129 12:20:11.878895 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:12 crc kubenswrapper[4993]: I0129 12:20:12.096003 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:12 crc kubenswrapper[4993]: W0129 12:20:12.107564 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c4c6d0f_3bc1_4b5c_a79a_5df8f9b35f69.slice/crio-8e2beaaff11120c54ec802fe65b24733682201bcb1cd1eaa68d2dfb3668da1a1 WatchSource:0}: Error finding container 8e2beaaff11120c54ec802fe65b24733682201bcb1cd1eaa68d2dfb3668da1a1: Status 404 returned error can't find the container with id 8e2beaaff11120c54ec802fe65b24733682201bcb1cd1eaa68d2dfb3668da1a1 Jan 29 12:20:12 crc kubenswrapper[4993]: I0129 12:20:12.755447 4993 generic.go:334] "Generic (PLEG): container finished" podID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerID="f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a" exitCode=0 Jan 29 12:20:12 crc kubenswrapper[4993]: I0129 12:20:12.755565 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerDied","Data":"f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a"} Jan 29 12:20:12 crc kubenswrapper[4993]: I0129 12:20:12.755843 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerStarted","Data":"8e2beaaff11120c54ec802fe65b24733682201bcb1cd1eaa68d2dfb3668da1a1"} Jan 29 12:20:14 crc kubenswrapper[4993]: I0129 12:20:14.770498 4993 generic.go:334] "Generic (PLEG): container finished" podID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerID="92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41" exitCode=0 Jan 29 12:20:14 crc kubenswrapper[4993]: I0129 12:20:14.770584 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerDied","Data":"92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41"} Jan 29 12:20:15 crc kubenswrapper[4993]: I0129 12:20:15.777442 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerStarted","Data":"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13"} Jan 29 12:20:18 crc kubenswrapper[4993]: I0129 12:20:18.156803 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:20:18 crc kubenswrapper[4993]: I0129 12:20:18.157979 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:20:21 crc kubenswrapper[4993]: I0129 12:20:21.879475 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:21 crc kubenswrapper[4993]: I0129 12:20:21.880140 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:21 crc kubenswrapper[4993]: I0129 12:20:21.919955 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:21 crc kubenswrapper[4993]: I0129 12:20:21.941247 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sl7zz" podStartSLOduration=8.47967734 podStartE2EDuration="10.941228669s" podCreationTimestamp="2026-01-29 12:20:11 +0000 UTC" firstStartedPulling="2026-01-29 12:20:12.757359206 +0000 UTC m=+896.772489332" lastFinishedPulling="2026-01-29 12:20:15.218910535 +0000 UTC m=+899.234040661" observedRunningTime="2026-01-29 12:20:15.802559265 +0000 UTC m=+899.817689391" watchObservedRunningTime="2026-01-29 12:20:21.941228669 +0000 UTC m=+905.956358805" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.798861 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq"] Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.800143 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.804158 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.806420 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq"] Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.867682 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.966048 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.966120 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:22 crc kubenswrapper[4993]: I0129 12:20:22.966222 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbfvv\" (UniqueName: \"kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.067606 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbfvv\" (UniqueName: \"kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.067689 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.067742 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.068326 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.068541 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.098002 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbfvv\" (UniqueName: \"kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.131603 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.339607 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq"] Jan 29 12:20:23 crc kubenswrapper[4993]: W0129 12:20:23.348552 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8952af7_d5e0_4a95_b584_347ba5d87a16.slice/crio-ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31 WatchSource:0}: Error finding container ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31: Status 404 returned error can't find the container with id ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31 Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.825877 4993 generic.go:334] "Generic (PLEG): container finished" podID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerID="64a566d964c9dbd399005de9e09a1472bac069b1385caf147cd9a14ffea34f9d" exitCode=0 Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.825968 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" event={"ID":"c8952af7-d5e0-4a95-b584-347ba5d87a16","Type":"ContainerDied","Data":"64a566d964c9dbd399005de9e09a1472bac069b1385caf147cd9a14ffea34f9d"} Jan 29 12:20:23 crc kubenswrapper[4993]: I0129 12:20:23.826262 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" event={"ID":"c8952af7-d5e0-4a95-b584-347ba5d87a16","Type":"ContainerStarted","Data":"ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31"} Jan 29 12:20:24 crc kubenswrapper[4993]: I0129 12:20:24.779819 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pbdv4" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" containerID="cri-o://8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9" gracePeriod=15 Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.647079 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pbdv4_4c84abb7-3adc-4707-8aae-c2e8add2ad05/console/0.log" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.647480 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711471 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711758 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711820 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711857 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711913 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711938 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwwv6\" (UniqueName: \"kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.711978 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert\") pod \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\" (UID: \"4c84abb7-3adc-4707-8aae-c2e8add2ad05\") " Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.712850 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.712712 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config" (OuterVolumeSpecName: "console-config") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.713251 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.713494 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca" (OuterVolumeSpecName: "service-ca") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.717783 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6" (OuterVolumeSpecName: "kube-api-access-nwwv6") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "kube-api-access-nwwv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.718014 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.719566 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4c84abb7-3adc-4707-8aae-c2e8add2ad05" (UID: "4c84abb7-3adc-4707-8aae-c2e8add2ad05"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812807 4993 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812839 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwwv6\" (UniqueName: \"kubernetes.io/projected/4c84abb7-3adc-4707-8aae-c2e8add2ad05-kube-api-access-nwwv6\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812849 4993 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812858 4993 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812867 4993 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812875 4993 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.812882 4993 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c84abb7-3adc-4707-8aae-c2e8add2ad05-console-config\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.841304 4993 generic.go:334] "Generic (PLEG): container finished" podID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerID="adbbafdf4b8e382952bf8d638cb10ca77fdb3e08b1533d61e87928aa41677575" exitCode=0 Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.841394 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" event={"ID":"c8952af7-d5e0-4a95-b584-347ba5d87a16","Type":"ContainerDied","Data":"adbbafdf4b8e382952bf8d638cb10ca77fdb3e08b1533d61e87928aa41677575"} Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843418 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pbdv4_4c84abb7-3adc-4707-8aae-c2e8add2ad05/console/0.log" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843475 4993 generic.go:334] "Generic (PLEG): container finished" podID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerID="8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9" exitCode=2 Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843509 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pbdv4" event={"ID":"4c84abb7-3adc-4707-8aae-c2e8add2ad05","Type":"ContainerDied","Data":"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9"} Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843542 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pbdv4" event={"ID":"4c84abb7-3adc-4707-8aae-c2e8add2ad05","Type":"ContainerDied","Data":"7c1e1b30d81f318073020956484aa23b562d8c78d1f07cdea8e85dd5531cb049"} Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843564 4993 scope.go:117] "RemoveContainer" containerID="8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.843564 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pbdv4" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.883538 4993 scope.go:117] "RemoveContainer" containerID="8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.883705 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:20:25 crc kubenswrapper[4993]: E0129 12:20:25.884103 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9\": container with ID starting with 8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9 not found: ID does not exist" containerID="8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.884150 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9"} err="failed to get container status \"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9\": rpc error: code = NotFound desc = could not find container \"8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9\": container with ID starting with 8a50dbd99da657d53646443aeb3fd0d452c58982757a3a6709c1209f31f30ff9 not found: ID does not exist" Jan 29 12:20:25 crc kubenswrapper[4993]: I0129 12:20:25.886811 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pbdv4"] Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.147659 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.148352 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sl7zz" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="registry-server" containerID="cri-o://8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13" gracePeriod=2 Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.487776 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.519621 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content\") pod \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.519761 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities\") pod \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.519788 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dmp8\" (UniqueName: \"kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8\") pod \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\" (UID: \"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69\") " Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.521127 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities" (OuterVolumeSpecName: "utilities") pod "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" (UID: "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.523878 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8" (OuterVolumeSpecName: "kube-api-access-2dmp8") pod "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" (UID: "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69"). InnerVolumeSpecName "kube-api-access-2dmp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.546089 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" (UID: "7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.621066 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.621099 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dmp8\" (UniqueName: \"kubernetes.io/projected/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-kube-api-access-2dmp8\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.621112 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.854015 4993 generic.go:334] "Generic (PLEG): container finished" podID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerID="8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13" exitCode=0 Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.854076 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerDied","Data":"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13"} Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.854124 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl7zz" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.854152 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl7zz" event={"ID":"7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69","Type":"ContainerDied","Data":"8e2beaaff11120c54ec802fe65b24733682201bcb1cd1eaa68d2dfb3668da1a1"} Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.854219 4993 scope.go:117] "RemoveContainer" containerID="8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.857499 4993 generic.go:334] "Generic (PLEG): container finished" podID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerID="6f733cfa1ce1a7519905605b6b668daa09e95b89d52ec1d244736a27ee6a4e84" exitCode=0 Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.857652 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" event={"ID":"c8952af7-d5e0-4a95-b584-347ba5d87a16","Type":"ContainerDied","Data":"6f733cfa1ce1a7519905605b6b668daa09e95b89d52ec1d244736a27ee6a4e84"} Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.872370 4993 scope.go:117] "RemoveContainer" containerID="92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.915755 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.920158 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl7zz"] Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.923759 4993 scope.go:117] "RemoveContainer" containerID="f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.937775 4993 scope.go:117] "RemoveContainer" containerID="8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13" Jan 29 12:20:26 crc kubenswrapper[4993]: E0129 12:20:26.940713 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13\": container with ID starting with 8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13 not found: ID does not exist" containerID="8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.940751 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13"} err="failed to get container status \"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13\": rpc error: code = NotFound desc = could not find container \"8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13\": container with ID starting with 8f117a0d55844246674221332ebb9f4ddd3f29293fa461e154cda67010418f13 not found: ID does not exist" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.940779 4993 scope.go:117] "RemoveContainer" containerID="92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41" Jan 29 12:20:26 crc kubenswrapper[4993]: E0129 12:20:26.941407 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41\": container with ID starting with 92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41 not found: ID does not exist" containerID="92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.941429 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41"} err="failed to get container status \"92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41\": rpc error: code = NotFound desc = could not find container \"92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41\": container with ID starting with 92df1962d5fb6297632f6eb538a9cb4a2ec0414d71dae006e5f95cf2ff573f41 not found: ID does not exist" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.941446 4993 scope.go:117] "RemoveContainer" containerID="f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a" Jan 29 12:20:26 crc kubenswrapper[4993]: E0129 12:20:26.941731 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a\": container with ID starting with f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a not found: ID does not exist" containerID="f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a" Jan 29 12:20:26 crc kubenswrapper[4993]: I0129 12:20:26.941751 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a"} err="failed to get container status \"f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a\": rpc error: code = NotFound desc = could not find container \"f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a\": container with ID starting with f8e542cacbf5ca318f5283c53b900b73f649dd44d7c137b3be064095f7611a7a not found: ID does not exist" Jan 29 12:20:27 crc kubenswrapper[4993]: I0129 12:20:27.201142 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" path="/var/lib/kubelet/pods/4c84abb7-3adc-4707-8aae-c2e8add2ad05/volumes" Jan 29 12:20:27 crc kubenswrapper[4993]: I0129 12:20:27.202170 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" path="/var/lib/kubelet/pods/7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69/volumes" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.083302 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.237495 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbfvv\" (UniqueName: \"kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv\") pod \"c8952af7-d5e0-4a95-b584-347ba5d87a16\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.237551 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle\") pod \"c8952af7-d5e0-4a95-b584-347ba5d87a16\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.237619 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util\") pod \"c8952af7-d5e0-4a95-b584-347ba5d87a16\" (UID: \"c8952af7-d5e0-4a95-b584-347ba5d87a16\") " Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.239033 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle" (OuterVolumeSpecName: "bundle") pod "c8952af7-d5e0-4a95-b584-347ba5d87a16" (UID: "c8952af7-d5e0-4a95-b584-347ba5d87a16"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.243726 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv" (OuterVolumeSpecName: "kube-api-access-mbfvv") pod "c8952af7-d5e0-4a95-b584-347ba5d87a16" (UID: "c8952af7-d5e0-4a95-b584-347ba5d87a16"). InnerVolumeSpecName "kube-api-access-mbfvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.252307 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util" (OuterVolumeSpecName: "util") pod "c8952af7-d5e0-4a95-b584-347ba5d87a16" (UID: "c8952af7-d5e0-4a95-b584-347ba5d87a16"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.339157 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbfvv\" (UniqueName: \"kubernetes.io/projected/c8952af7-d5e0-4a95-b584-347ba5d87a16-kube-api-access-mbfvv\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.339233 4993 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.339247 4993 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c8952af7-d5e0-4a95-b584-347ba5d87a16-util\") on node \"crc\" DevicePath \"\"" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.874094 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" event={"ID":"c8952af7-d5e0-4a95-b584-347ba5d87a16","Type":"ContainerDied","Data":"ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31"} Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.874138 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae7e28050ea9a0cb84ecf0929fa85331419e5bc6ccfca2f3c6d3eb3754fcbe31" Jan 29 12:20:28 crc kubenswrapper[4993]: I0129 12:20:28.874239 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.995339 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5496d86f-5p52h"] Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996119 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="pull" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996133 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="pull" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996144 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="util" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996149 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="util" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996157 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996163 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996174 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="extract-utilities" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996196 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="extract-utilities" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996206 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="registry-server" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996211 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="registry-server" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996222 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="extract" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996228 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="extract" Jan 29 12:20:38 crc kubenswrapper[4993]: E0129 12:20:38.996239 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="extract-content" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996244 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="extract-content" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996348 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8952af7-d5e0-4a95-b584-347ba5d87a16" containerName="extract" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996361 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4c6d0f-3bc1-4b5c-a79a-5df8f9b35f69" containerName="registry-server" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996372 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c84abb7-3adc-4707-8aae-c2e8add2ad05" containerName="console" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.996768 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.999660 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 29 12:20:38 crc kubenswrapper[4993]: I0129 12:20:38.999675 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.000805 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.001064 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-d986d" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.002313 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.023561 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5496d86f-5p52h"] Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.165510 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-apiservice-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.165597 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-webhook-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.165629 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gnl\" (UniqueName: \"kubernetes.io/projected/a7d00ccb-f451-43bb-8280-08604b5ddf86-kube-api-access-m8gnl\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.266608 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-apiservice-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.266677 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-webhook-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.266702 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gnl\" (UniqueName: \"kubernetes.io/projected/a7d00ccb-f451-43bb-8280-08604b5ddf86-kube-api-access-m8gnl\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.290150 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-webhook-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.290220 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d00ccb-f451-43bb-8280-08604b5ddf86-apiservice-cert\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.301032 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gnl\" (UniqueName: \"kubernetes.io/projected/a7d00ccb-f451-43bb-8280-08604b5ddf86-kube-api-access-m8gnl\") pod \"metallb-operator-controller-manager-5496d86f-5p52h\" (UID: \"a7d00ccb-f451-43bb-8280-08604b5ddf86\") " pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.310931 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.342006 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2"] Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.353175 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.363254 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.363448 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.364074 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sckh8" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.373692 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2"] Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.471136 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2xn\" (UniqueName: \"kubernetes.io/projected/89b47e24-436a-4914-b4ff-c9767a62ac58-kube-api-access-6p2xn\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.471357 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-webhook-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.471451 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-apiservice-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.573821 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2xn\" (UniqueName: \"kubernetes.io/projected/89b47e24-436a-4914-b4ff-c9767a62ac58-kube-api-access-6p2xn\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.573903 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-webhook-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.573933 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-apiservice-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.577875 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-apiservice-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.584238 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/89b47e24-436a-4914-b4ff-c9767a62ac58-webhook-cert\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.601069 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2xn\" (UniqueName: \"kubernetes.io/projected/89b47e24-436a-4914-b4ff-c9767a62ac58-kube-api-access-6p2xn\") pod \"metallb-operator-webhook-server-75647f5f8d-xh6r2\" (UID: \"89b47e24-436a-4914-b4ff-c9767a62ac58\") " pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.637071 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5496d86f-5p52h"] Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.694258 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.906903 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2"] Jan 29 12:20:39 crc kubenswrapper[4993]: W0129 12:20:39.912176 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89b47e24_436a_4914_b4ff_c9767a62ac58.slice/crio-21987cc5494af82451d539760ef149ad18f0d1fdfd75d4ae3eab011e9b827661 WatchSource:0}: Error finding container 21987cc5494af82451d539760ef149ad18f0d1fdfd75d4ae3eab011e9b827661: Status 404 returned error can't find the container with id 21987cc5494af82451d539760ef149ad18f0d1fdfd75d4ae3eab011e9b827661 Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.929929 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" event={"ID":"a7d00ccb-f451-43bb-8280-08604b5ddf86","Type":"ContainerStarted","Data":"9e028cfeb812c439178b78be0409d45f8275e6f16efca0d887735ce95eb8e5dd"} Jan 29 12:20:39 crc kubenswrapper[4993]: I0129 12:20:39.930894 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" event={"ID":"89b47e24-436a-4914-b4ff-c9767a62ac58","Type":"ContainerStarted","Data":"21987cc5494af82451d539760ef149ad18f0d1fdfd75d4ae3eab011e9b827661"} Jan 29 12:20:42 crc kubenswrapper[4993]: I0129 12:20:42.953230 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" event={"ID":"a7d00ccb-f451-43bb-8280-08604b5ddf86","Type":"ContainerStarted","Data":"78e762cd0ff479d276067e8876e86f7ef0f0c4f5299644e274ad929b2718f0fd"} Jan 29 12:20:42 crc kubenswrapper[4993]: I0129 12:20:42.953817 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:20:42 crc kubenswrapper[4993]: I0129 12:20:42.977286 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" podStartSLOduration=2.143016747 podStartE2EDuration="4.97726312s" podCreationTimestamp="2026-01-29 12:20:38 +0000 UTC" firstStartedPulling="2026-01-29 12:20:39.639367028 +0000 UTC m=+923.654497154" lastFinishedPulling="2026-01-29 12:20:42.473613401 +0000 UTC m=+926.488743527" observedRunningTime="2026-01-29 12:20:42.972866254 +0000 UTC m=+926.987996390" watchObservedRunningTime="2026-01-29 12:20:42.97726312 +0000 UTC m=+926.992393256" Jan 29 12:20:44 crc kubenswrapper[4993]: I0129 12:20:44.968374 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" event={"ID":"89b47e24-436a-4914-b4ff-c9767a62ac58","Type":"ContainerStarted","Data":"1a7cc418cdcb5cb297bca37f11a0211605be9089d7ea663a4da60be533b12042"} Jan 29 12:20:44 crc kubenswrapper[4993]: I0129 12:20:44.969009 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:20:44 crc kubenswrapper[4993]: I0129 12:20:44.990872 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" podStartSLOduration=1.483032804 podStartE2EDuration="5.990849967s" podCreationTimestamp="2026-01-29 12:20:39 +0000 UTC" firstStartedPulling="2026-01-29 12:20:39.915600074 +0000 UTC m=+923.930730200" lastFinishedPulling="2026-01-29 12:20:44.423417237 +0000 UTC m=+928.438547363" observedRunningTime="2026-01-29 12:20:44.985443633 +0000 UTC m=+929.000573759" watchObservedRunningTime="2026-01-29 12:20:44.990849967 +0000 UTC m=+929.005980103" Jan 29 12:20:48 crc kubenswrapper[4993]: I0129 12:20:48.156418 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:20:48 crc kubenswrapper[4993]: I0129 12:20:48.156685 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:20:59 crc kubenswrapper[4993]: I0129 12:20:59.698522 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-75647f5f8d-xh6r2" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.216642 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.218574 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.236217 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.328638 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4r9\" (UniqueName: \"kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.328696 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.328805 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.429821 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4r9\" (UniqueName: \"kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.429873 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.429894 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.430396 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.430521 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.449326 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4r9\" (UniqueName: \"kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9\") pod \"community-operators-cd8gj\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.534084 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:07 crc kubenswrapper[4993]: I0129 12:21:07.840674 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:08 crc kubenswrapper[4993]: I0129 12:21:08.093989 4993 generic.go:334] "Generic (PLEG): container finished" podID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerID="6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3" exitCode=0 Jan 29 12:21:08 crc kubenswrapper[4993]: I0129 12:21:08.094259 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerDied","Data":"6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3"} Jan 29 12:21:08 crc kubenswrapper[4993]: I0129 12:21:08.094328 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerStarted","Data":"f2229582232639deae75be030ff7b93e1433bea5812256ae114c55bcd8ceb706"} Jan 29 12:21:10 crc kubenswrapper[4993]: I0129 12:21:10.105462 4993 generic.go:334] "Generic (PLEG): container finished" podID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerID="4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86" exitCode=0 Jan 29 12:21:10 crc kubenswrapper[4993]: I0129 12:21:10.105550 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerDied","Data":"4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86"} Jan 29 12:21:11 crc kubenswrapper[4993]: I0129 12:21:11.118556 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerStarted","Data":"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156"} Jan 29 12:21:11 crc kubenswrapper[4993]: I0129 12:21:11.141686 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cd8gj" podStartSLOduration=1.720800213 podStartE2EDuration="4.141666313s" podCreationTimestamp="2026-01-29 12:21:07 +0000 UTC" firstStartedPulling="2026-01-29 12:21:08.095710724 +0000 UTC m=+952.110840850" lastFinishedPulling="2026-01-29 12:21:10.516576824 +0000 UTC m=+954.531706950" observedRunningTime="2026-01-29 12:21:11.140229155 +0000 UTC m=+955.155359301" watchObservedRunningTime="2026-01-29 12:21:11.141666313 +0000 UTC m=+955.156796439" Jan 29 12:21:17 crc kubenswrapper[4993]: I0129 12:21:17.534372 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:17 crc kubenswrapper[4993]: I0129 12:21:17.534815 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:17 crc kubenswrapper[4993]: I0129 12:21:17.591222 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.156224 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.156298 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.156353 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.157327 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.157407 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141" gracePeriod=600 Jan 29 12:21:18 crc kubenswrapper[4993]: I0129 12:21:18.199220 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.165799 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141" exitCode=0 Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.165894 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141"} Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.166461 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47"} Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.166496 4993 scope.go:117] "RemoveContainer" containerID="1d911c3fa471b0db4f6ebcc68988ce7072f59a22228381a7f5b2aca4ee2fbffd" Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.313749 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5496d86f-5p52h" Jan 29 12:21:19 crc kubenswrapper[4993]: I0129 12:21:19.994488 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.016693 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sjhwc"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.030920 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.036846 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.038045 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.038050 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.038357 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7sk8x" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.049540 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.050078 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.057171 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.122052 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-99rlj"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.123242 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.126943 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.128903 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.128971 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-99fq9" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.128905 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.134453 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-zlzsh"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.136154 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.137784 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.141441 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-zlzsh"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.174936 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cd8gj" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="registry-server" containerID="cri-o://9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156" gracePeriod=2 Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.194931 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-conf\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.194986 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195009 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-sockets\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195034 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195051 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-startup\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195079 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-reloader\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195103 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njdhw\" (UniqueName: \"kubernetes.io/projected/ec787c89-bd58-470e-942e-b9775f1ee08e-kube-api-access-njdhw\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195221 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxb2z\" (UniqueName: \"kubernetes.io/projected/8f6aec89-4069-4b04-8258-d9fbeb62bd79-kube-api-access-nxb2z\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.195242 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics-certs\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.296890 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-reloader\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297230 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njdhw\" (UniqueName: \"kubernetes.io/projected/ec787c89-bd58-470e-942e-b9775f1ee08e-kube-api-access-njdhw\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297356 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-metrics-certs\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297515 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-cert\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297527 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-reloader\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297841 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxb2z\" (UniqueName: \"kubernetes.io/projected/8f6aec89-4069-4b04-8258-d9fbeb62bd79-kube-api-access-nxb2z\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297916 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics-certs\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297962 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-conf\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.297991 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c25e8c1-f850-435e-956c-92cb3e13be64-metallb-excludel2\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298023 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qlwh\" (UniqueName: \"kubernetes.io/projected/36e0eaff-f539-449a-8cab-3a8e67a503da-kube-api-access-5qlwh\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298051 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298084 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z8sq\" (UniqueName: \"kubernetes.io/projected/5c25e8c1-f850-435e-956c-92cb3e13be64-kube-api-access-7z8sq\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298109 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-sockets\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298138 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-metrics-certs\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298160 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298202 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-startup\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298230 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.298416 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-conf\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.298754 4993 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.298817 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert podName:8f6aec89-4069-4b04-8258-d9fbeb62bd79 nodeName:}" failed. No retries permitted until 2026-01-29 12:21:20.79880092 +0000 UTC m=+964.813931046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert") pod "frr-k8s-webhook-server-7df86c4f6c-m5726" (UID: "8f6aec89-4069-4b04-8258-d9fbeb62bd79") : secret "frr-k8s-webhook-server-cert" not found Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.299062 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-sockets\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.299154 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.299942 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec787c89-bd58-470e-942e-b9775f1ee08e-frr-startup\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.304696 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec787c89-bd58-470e-942e-b9775f1ee08e-metrics-certs\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.317685 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njdhw\" (UniqueName: \"kubernetes.io/projected/ec787c89-bd58-470e-942e-b9775f1ee08e-kube-api-access-njdhw\") pod \"frr-k8s-sjhwc\" (UID: \"ec787c89-bd58-470e-942e-b9775f1ee08e\") " pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.331495 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxb2z\" (UniqueName: \"kubernetes.io/projected/8f6aec89-4069-4b04-8258-d9fbeb62bd79-kube-api-access-nxb2z\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.379735 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.398771 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.398864 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-metrics-certs\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.398893 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-cert\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.399842 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c25e8c1-f850-435e-956c-92cb3e13be64-metallb-excludel2\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.399876 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qlwh\" (UniqueName: \"kubernetes.io/projected/36e0eaff-f539-449a-8cab-3a8e67a503da-kube-api-access-5qlwh\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.399899 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z8sq\" (UniqueName: \"kubernetes.io/projected/5c25e8c1-f850-435e-956c-92cb3e13be64-kube-api-access-7z8sq\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.399928 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-metrics-certs\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.399283 4993 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.400315 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist podName:5c25e8c1-f850-435e-956c-92cb3e13be64 nodeName:}" failed. No retries permitted until 2026-01-29 12:21:20.900280182 +0000 UTC m=+964.915410378 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist") pod "speaker-99rlj" (UID: "5c25e8c1-f850-435e-956c-92cb3e13be64") : secret "metallb-memberlist" not found Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.400784 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c25e8c1-f850-435e-956c-92cb3e13be64-metallb-excludel2\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.404417 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-metrics-certs\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.405076 4993 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.405546 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-metrics-certs\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.411917 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36e0eaff-f539-449a-8cab-3a8e67a503da-cert\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.417879 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z8sq\" (UniqueName: \"kubernetes.io/projected/5c25e8c1-f850-435e-956c-92cb3e13be64-kube-api-access-7z8sq\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.418992 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qlwh\" (UniqueName: \"kubernetes.io/projected/36e0eaff-f539-449a-8cab-3a8e67a503da-kube-api-access-5qlwh\") pod \"controller-6968d8fdc4-zlzsh\" (UID: \"36e0eaff-f539-449a-8cab-3a8e67a503da\") " pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.448445 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.805104 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.811153 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8f6aec89-4069-4b04-8258-d9fbeb62bd79-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m5726\" (UID: \"8f6aec89-4069-4b04-8258-d9fbeb62bd79\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.837218 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-zlzsh"] Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.906123 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.906350 4993 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 29 12:21:20 crc kubenswrapper[4993]: E0129 12:21:20.906441 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist podName:5c25e8c1-f850-435e-956c-92cb3e13be64 nodeName:}" failed. No retries permitted until 2026-01-29 12:21:21.906417537 +0000 UTC m=+965.921547663 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist") pod "speaker-99rlj" (UID: "5c25e8c1-f850-435e-956c-92cb3e13be64") : secret "metallb-memberlist" not found Jan 29 12:21:20 crc kubenswrapper[4993]: I0129 12:21:20.993758 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:21 crc kubenswrapper[4993]: I0129 12:21:21.181872 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-zlzsh" event={"ID":"36e0eaff-f539-449a-8cab-3a8e67a503da","Type":"ContainerStarted","Data":"5a487d3cba4941866c251eabf4f60d130adc2e63d73afbc26559add5e943faca"} Jan 29 12:21:21 crc kubenswrapper[4993]: W0129 12:21:21.202912 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f6aec89_4069_4b04_8258_d9fbeb62bd79.slice/crio-3618cb36f6af55aedf4214c1111403cb58bc9b71a2b0868c573430b408965740 WatchSource:0}: Error finding container 3618cb36f6af55aedf4214c1111403cb58bc9b71a2b0868c573430b408965740: Status 404 returned error can't find the container with id 3618cb36f6af55aedf4214c1111403cb58bc9b71a2b0868c573430b408965740 Jan 29 12:21:21 crc kubenswrapper[4993]: I0129 12:21:21.203940 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726"] Jan 29 12:21:21 crc kubenswrapper[4993]: I0129 12:21:21.919786 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:21 crc kubenswrapper[4993]: I0129 12:21:21.941549 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c25e8c1-f850-435e-956c-92cb3e13be64-memberlist\") pod \"speaker-99rlj\" (UID: \"5c25e8c1-f850-435e-956c-92cb3e13be64\") " pod="metallb-system/speaker-99rlj" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.036950 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.197683 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"89c32c240b63cd3132b64ad1c51d457c15d933a8f23bcc113973d8e0b096c318"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.199291 4993 generic.go:334] "Generic (PLEG): container finished" podID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerID="9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156" exitCode=0 Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.199362 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerDied","Data":"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.199382 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cd8gj" event={"ID":"b9f98c7d-a3b2-459a-9b17-c82496480d6f","Type":"ContainerDied","Data":"f2229582232639deae75be030ff7b93e1433bea5812256ae114c55bcd8ceb706"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.199396 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cd8gj" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.199404 4993 scope.go:117] "RemoveContainer" containerID="9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.202072 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" event={"ID":"8f6aec89-4069-4b04-8258-d9fbeb62bd79","Type":"ContainerStarted","Data":"3618cb36f6af55aedf4214c1111403cb58bc9b71a2b0868c573430b408965740"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.211244 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-zlzsh" event={"ID":"36e0eaff-f539-449a-8cab-3a8e67a503da","Type":"ContainerStarted","Data":"99129eaf1ccdb80ade0826fb437c19a49675f038823fd91d3a05dc391bd02768"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.211345 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-zlzsh" event={"ID":"36e0eaff-f539-449a-8cab-3a8e67a503da","Type":"ContainerStarted","Data":"40f6d276d5ff7a9ae5de4a7b3cae7c5f3c172c723698dc0c484973c1acb673a5"} Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.211557 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.220459 4993 scope.go:117] "RemoveContainer" containerID="4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.223665 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4r9\" (UniqueName: \"kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9\") pod \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.223701 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities\") pod \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.223736 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content\") pod \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\" (UID: \"b9f98c7d-a3b2-459a-9b17-c82496480d6f\") " Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.224662 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities" (OuterVolumeSpecName: "utilities") pod "b9f98c7d-a3b2-459a-9b17-c82496480d6f" (UID: "b9f98c7d-a3b2-459a-9b17-c82496480d6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.229678 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9" (OuterVolumeSpecName: "kube-api-access-qd4r9") pod "b9f98c7d-a3b2-459a-9b17-c82496480d6f" (UID: "b9f98c7d-a3b2-459a-9b17-c82496480d6f"). InnerVolumeSpecName "kube-api-access-qd4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.235573 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-zlzsh" podStartSLOduration=2.23555154 podStartE2EDuration="2.23555154s" podCreationTimestamp="2026-01-29 12:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:21:22.230160567 +0000 UTC m=+966.245290713" watchObservedRunningTime="2026-01-29 12:21:22.23555154 +0000 UTC m=+966.250681676" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.239168 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-99rlj" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.255812 4993 scope.go:117] "RemoveContainer" containerID="6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.287913 4993 scope.go:117] "RemoveContainer" containerID="9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156" Jan 29 12:21:22 crc kubenswrapper[4993]: E0129 12:21:22.288829 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156\": container with ID starting with 9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156 not found: ID does not exist" containerID="9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.288866 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156"} err="failed to get container status \"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156\": rpc error: code = NotFound desc = could not find container \"9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156\": container with ID starting with 9c0b9e2907c86f91c23126c1b6ee860c4f28ef664261967be3f0663398f8f156 not found: ID does not exist" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.288888 4993 scope.go:117] "RemoveContainer" containerID="4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86" Jan 29 12:21:22 crc kubenswrapper[4993]: E0129 12:21:22.290203 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86\": container with ID starting with 4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86 not found: ID does not exist" containerID="4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.290238 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86"} err="failed to get container status \"4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86\": rpc error: code = NotFound desc = could not find container \"4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86\": container with ID starting with 4c238b42053c292276d80b2e73c465e306f203ec636fb99036254a2c288f5e86 not found: ID does not exist" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.290258 4993 scope.go:117] "RemoveContainer" containerID="6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3" Jan 29 12:21:22 crc kubenswrapper[4993]: E0129 12:21:22.290744 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3\": container with ID starting with 6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3 not found: ID does not exist" containerID="6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.290829 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3"} err="failed to get container status \"6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3\": rpc error: code = NotFound desc = could not find container \"6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3\": container with ID starting with 6dd390a0a034e7eb430f4f862d7f9e11644320ae42c954d72459ab141b5ae4f3 not found: ID does not exist" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.292858 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9f98c7d-a3b2-459a-9b17-c82496480d6f" (UID: "b9f98c7d-a3b2-459a-9b17-c82496480d6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.325109 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4r9\" (UniqueName: \"kubernetes.io/projected/b9f98c7d-a3b2-459a-9b17-c82496480d6f-kube-api-access-qd4r9\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.325867 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.325918 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9f98c7d-a3b2-459a-9b17-c82496480d6f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.532566 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:22 crc kubenswrapper[4993]: I0129 12:21:22.541429 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cd8gj"] Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.208686 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" path="/var/lib/kubelet/pods/b9f98c7d-a3b2-459a-9b17-c82496480d6f/volumes" Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.235262 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-99rlj" event={"ID":"5c25e8c1-f850-435e-956c-92cb3e13be64","Type":"ContainerStarted","Data":"856ad66c2b0b6edb57031cd5dfa7226d8082361b0b4e57b65f05edb447d550df"} Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.235303 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-99rlj" event={"ID":"5c25e8c1-f850-435e-956c-92cb3e13be64","Type":"ContainerStarted","Data":"b727389fc7468b65cf011e1d4a8cc2a964d24f21e67f5daf8b44856c99d4c91f"} Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.235314 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-99rlj" event={"ID":"5c25e8c1-f850-435e-956c-92cb3e13be64","Type":"ContainerStarted","Data":"57fa37e775822a0f9132fe1bdfd852bb8be52d045580e302a4fbc4149ba560d3"} Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.235426 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-99rlj" Jan 29 12:21:23 crc kubenswrapper[4993]: I0129 12:21:23.258553 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-99rlj" podStartSLOduration=3.2585304219999998 podStartE2EDuration="3.258530422s" podCreationTimestamp="2026-01-29 12:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:21:23.253356156 +0000 UTC m=+967.268486282" watchObservedRunningTime="2026-01-29 12:21:23.258530422 +0000 UTC m=+967.273660548" Jan 29 12:21:31 crc kubenswrapper[4993]: I0129 12:21:31.287220 4993 generic.go:334] "Generic (PLEG): container finished" podID="ec787c89-bd58-470e-942e-b9775f1ee08e" containerID="99b26448f3fe1b558351b48cf649b927391388b7329d2400bd03ba3eee45da25" exitCode=0 Jan 29 12:21:31 crc kubenswrapper[4993]: I0129 12:21:31.287300 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerDied","Data":"99b26448f3fe1b558351b48cf649b927391388b7329d2400bd03ba3eee45da25"} Jan 29 12:21:31 crc kubenswrapper[4993]: I0129 12:21:31.292303 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" event={"ID":"8f6aec89-4069-4b04-8258-d9fbeb62bd79","Type":"ContainerStarted","Data":"10d980d0da25a1ed8c07f43e757c225638c78f1d02f4e5a38e5fa29e9842e815"} Jan 29 12:21:31 crc kubenswrapper[4993]: I0129 12:21:31.292578 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:31 crc kubenswrapper[4993]: I0129 12:21:31.337156 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" podStartSLOduration=3.042970579 podStartE2EDuration="12.337131021s" podCreationTimestamp="2026-01-29 12:21:19 +0000 UTC" firstStartedPulling="2026-01-29 12:21:21.205304625 +0000 UTC m=+965.220434751" lastFinishedPulling="2026-01-29 12:21:30.499465067 +0000 UTC m=+974.514595193" observedRunningTime="2026-01-29 12:21:31.33407259 +0000 UTC m=+975.349202716" watchObservedRunningTime="2026-01-29 12:21:31.337131021 +0000 UTC m=+975.352261157" Jan 29 12:21:32 crc kubenswrapper[4993]: I0129 12:21:32.243097 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-99rlj" Jan 29 12:21:32 crc kubenswrapper[4993]: I0129 12:21:32.299543 4993 generic.go:334] "Generic (PLEG): container finished" podID="ec787c89-bd58-470e-942e-b9775f1ee08e" containerID="d7e74556cd5785445f4e5a120f80f7ccfce7bab8e30be7123248feeb3e5f5c50" exitCode=0 Jan 29 12:21:32 crc kubenswrapper[4993]: I0129 12:21:32.299652 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerDied","Data":"d7e74556cd5785445f4e5a120f80f7ccfce7bab8e30be7123248feeb3e5f5c50"} Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.308265 4993 generic.go:334] "Generic (PLEG): container finished" podID="ec787c89-bd58-470e-942e-b9775f1ee08e" containerID="03127d0312f7fb8463d3812df6286f6f37c6d6d34b49cf5e67121cb519732508" exitCode=0 Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.308314 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerDied","Data":"03127d0312f7fb8463d3812df6286f6f37c6d6d34b49cf5e67121cb519732508"} Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.613714 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl"] Jan 29 12:21:33 crc kubenswrapper[4993]: E0129 12:21:33.614359 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="extract-content" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.614380 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="extract-content" Jan 29 12:21:33 crc kubenswrapper[4993]: E0129 12:21:33.614392 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="registry-server" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.614399 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="registry-server" Jan 29 12:21:33 crc kubenswrapper[4993]: E0129 12:21:33.614413 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="extract-utilities" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.614421 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="extract-utilities" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.614547 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f98c7d-a3b2-459a-9b17-c82496480d6f" containerName="registry-server" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.615467 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.619294 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.627217 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl"] Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.791041 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.791120 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghzs\" (UniqueName: \"kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.791542 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.892752 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.892854 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.892892 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghzs\" (UniqueName: \"kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.893732 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.893887 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.921044 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghzs\" (UniqueName: \"kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:33 crc kubenswrapper[4993]: I0129 12:21:33.931018 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:34 crc kubenswrapper[4993]: I0129 12:21:34.319284 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"49b135ddb05b3b803053f46c96176de21e16ee03c75e69efb0e8edb0efe09183"} Jan 29 12:21:34 crc kubenswrapper[4993]: I0129 12:21:34.319820 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"bb45ca44c84db35191d9541d20ae9f983eef9daaa27304bfe9d78835ae919c7f"} Jan 29 12:21:34 crc kubenswrapper[4993]: I0129 12:21:34.364294 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl"] Jan 29 12:21:34 crc kubenswrapper[4993]: W0129 12:21:34.364867 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbc114ee_d9f7_4bdb_a104_6e8b1c7aaf6a.slice/crio-7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d WatchSource:0}: Error finding container 7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d: Status 404 returned error can't find the container with id 7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d Jan 29 12:21:35 crc kubenswrapper[4993]: I0129 12:21:35.327111 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"5cbcc6a3210b2788b22dd4875f8958e389736f43b57ede4c003078860a666ac1"} Jan 29 12:21:35 crc kubenswrapper[4993]: I0129 12:21:35.328316 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" event={"ID":"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a","Type":"ContainerStarted","Data":"7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d"} Jan 29 12:21:36 crc kubenswrapper[4993]: I0129 12:21:36.395311 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"cdcafb5e15e6dc7411c99282e003d40bee7a177d7f7ab0b89549fa4c947e8ce3"} Jan 29 12:21:36 crc kubenswrapper[4993]: I0129 12:21:36.401133 4993 generic.go:334] "Generic (PLEG): container finished" podID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerID="c7522eef942349fe235973825fb3690906d0383af73061062cb767c4c68d0885" exitCode=0 Jan 29 12:21:36 crc kubenswrapper[4993]: I0129 12:21:36.401172 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" event={"ID":"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a","Type":"ContainerDied","Data":"c7522eef942349fe235973825fb3690906d0383af73061062cb767c4c68d0885"} Jan 29 12:21:37 crc kubenswrapper[4993]: I0129 12:21:37.412029 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"c7be752a31c74c9868597e1288b9e99c73f3ef38e18750c7c01dbee2cc9b254f"} Jan 29 12:21:37 crc kubenswrapper[4993]: I0129 12:21:37.412621 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sjhwc" event={"ID":"ec787c89-bd58-470e-942e-b9775f1ee08e","Type":"ContainerStarted","Data":"ac22dacd3a86ad84c21f3e4dd046a2f0dfacc51c0c4baede32f974e9aae2f12f"} Jan 29 12:21:37 crc kubenswrapper[4993]: I0129 12:21:37.412656 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:37 crc kubenswrapper[4993]: I0129 12:21:37.436956 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sjhwc" podStartSLOduration=9.420126735 podStartE2EDuration="18.436927168s" podCreationTimestamp="2026-01-29 12:21:19 +0000 UTC" firstStartedPulling="2026-01-29 12:21:21.459639671 +0000 UTC m=+965.474769807" lastFinishedPulling="2026-01-29 12:21:30.476440114 +0000 UTC m=+974.491570240" observedRunningTime="2026-01-29 12:21:37.436791144 +0000 UTC m=+981.451921290" watchObservedRunningTime="2026-01-29 12:21:37.436927168 +0000 UTC m=+981.452057294" Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.380868 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.432962 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.433023 4993 generic.go:334] "Generic (PLEG): container finished" podID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerID="81dafababb0d4a9d00e4619f0162be1f3e79cafe76b2c49cad784f6566922430" exitCode=0 Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.433370 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" event={"ID":"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a","Type":"ContainerDied","Data":"81dafababb0d4a9d00e4619f0162be1f3e79cafe76b2c49cad784f6566922430"} Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.455296 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-zlzsh" Jan 29 12:21:40 crc kubenswrapper[4993]: I0129 12:21:40.999500 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m5726" Jan 29 12:21:41 crc kubenswrapper[4993]: I0129 12:21:41.440409 4993 generic.go:334] "Generic (PLEG): container finished" podID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerID="b9036401586f0d0745cbdded317cf275c52f7641154644cc6c3adfbeba63eeae" exitCode=0 Jan 29 12:21:41 crc kubenswrapper[4993]: I0129 12:21:41.440474 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" event={"ID":"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a","Type":"ContainerDied","Data":"b9036401586f0d0745cbdded317cf275c52f7641154644cc6c3adfbeba63eeae"} Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.688342 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.839349 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util\") pod \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.839432 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle\") pod \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.840346 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xghzs\" (UniqueName: \"kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs\") pod \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\" (UID: \"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a\") " Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.841147 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle" (OuterVolumeSpecName: "bundle") pod "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" (UID: "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.845874 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs" (OuterVolumeSpecName: "kube-api-access-xghzs") pod "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" (UID: "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a"). InnerVolumeSpecName "kube-api-access-xghzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.852004 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util" (OuterVolumeSpecName: "util") pod "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" (UID: "cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.941096 4993 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-util\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.941137 4993 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:42 crc kubenswrapper[4993]: I0129 12:21:42.941151 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xghzs\" (UniqueName: \"kubernetes.io/projected/cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a-kube-api-access-xghzs\") on node \"crc\" DevicePath \"\"" Jan 29 12:21:43 crc kubenswrapper[4993]: I0129 12:21:43.498579 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" event={"ID":"cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a","Type":"ContainerDied","Data":"7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d"} Jan 29 12:21:43 crc kubenswrapper[4993]: I0129 12:21:43.498980 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c756bf32423e47a33b2ad1bf055443df2254fcf3079756caa0435b78a58636d" Jan 29 12:21:43 crc kubenswrapper[4993]: I0129 12:21:43.498649 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.991626 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p"] Jan 29 12:21:46 crc kubenswrapper[4993]: E0129 12:21:46.992117 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="extract" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.992129 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="extract" Jan 29 12:21:46 crc kubenswrapper[4993]: E0129 12:21:46.992144 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="pull" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.992151 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="pull" Jan 29 12:21:46 crc kubenswrapper[4993]: E0129 12:21:46.992158 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="util" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.992165 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="util" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.992316 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a" containerName="extract" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.992810 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.994828 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.995047 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 29 12:21:46 crc kubenswrapper[4993]: I0129 12:21:46.995225 4993 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-zzgvq" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.020258 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p"] Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.090139 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.090282 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqqzr\" (UniqueName: \"kubernetes.io/projected/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-kube-api-access-dqqzr\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.191264 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqqzr\" (UniqueName: \"kubernetes.io/projected/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-kube-api-access-dqqzr\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.191368 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.191899 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.210141 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqqzr\" (UniqueName: \"kubernetes.io/projected/2bdcdda3-34ab-4a44-82fe-349a20d8d8d7-kube-api-access-dqqzr\") pod \"cert-manager-operator-controller-manager-66c8bdd694-ws28p\" (UID: \"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.307747 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" Jan 29 12:21:47 crc kubenswrapper[4993]: I0129 12:21:47.860856 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p"] Jan 29 12:21:47 crc kubenswrapper[4993]: W0129 12:21:47.863415 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bdcdda3_34ab_4a44_82fe_349a20d8d8d7.slice/crio-750e2a4d687a6136f3b05a8b7ff4da609ab77cd44eedd4b2f937cf8d6deb1bdc WatchSource:0}: Error finding container 750e2a4d687a6136f3b05a8b7ff4da609ab77cd44eedd4b2f937cf8d6deb1bdc: Status 404 returned error can't find the container with id 750e2a4d687a6136f3b05a8b7ff4da609ab77cd44eedd4b2f937cf8d6deb1bdc Jan 29 12:21:48 crc kubenswrapper[4993]: I0129 12:21:48.531271 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" event={"ID":"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7","Type":"ContainerStarted","Data":"750e2a4d687a6136f3b05a8b7ff4da609ab77cd44eedd4b2f937cf8d6deb1bdc"} Jan 29 12:21:50 crc kubenswrapper[4993]: I0129 12:21:50.382856 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sjhwc" Jan 29 12:21:52 crc kubenswrapper[4993]: I0129 12:21:52.581523 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" event={"ID":"2bdcdda3-34ab-4a44-82fe-349a20d8d8d7","Type":"ContainerStarted","Data":"fb6e8a7760db31985edbf26f0be3cf134ae151a148fe87bfa4079a34a8fc6688"} Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.586163 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-ws28p" podStartSLOduration=4.8212284709999995 podStartE2EDuration="8.586148144s" podCreationTimestamp="2026-01-29 12:21:46 +0000 UTC" firstStartedPulling="2026-01-29 12:21:47.866126231 +0000 UTC m=+991.881256357" lastFinishedPulling="2026-01-29 12:21:51.631045904 +0000 UTC m=+995.646176030" observedRunningTime="2026-01-29 12:21:52.611382727 +0000 UTC m=+996.626512863" watchObservedRunningTime="2026-01-29 12:21:54.586148144 +0000 UTC m=+998.601278270" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.589000 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-z9fkp"] Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.589721 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.591529 4993 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-46lmr" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.592024 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.593111 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.604031 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-z9fkp"] Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.688802 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-524zq\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-kube-api-access-524zq\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.688914 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.790275 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-524zq\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-kube-api-access-524zq\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.790365 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.846019 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.846831 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-524zq\" (UniqueName: \"kubernetes.io/projected/58863fed-2920-49fd-a1bc-6a5f8a771783-kube-api-access-524zq\") pod \"cert-manager-webhook-6888856db4-z9fkp\" (UID: \"58863fed-2920-49fd-a1bc-6a5f8a771783\") " pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:54 crc kubenswrapper[4993]: I0129 12:21:54.906846 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:21:55 crc kubenswrapper[4993]: I0129 12:21:55.550318 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-z9fkp"] Jan 29 12:21:55 crc kubenswrapper[4993]: W0129 12:21:55.559896 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58863fed_2920_49fd_a1bc_6a5f8a771783.slice/crio-e4ab74550d5065fc0f9e462ef70761f072a2ea48856618359761cc352b9ba0f8 WatchSource:0}: Error finding container e4ab74550d5065fc0f9e462ef70761f072a2ea48856618359761cc352b9ba0f8: Status 404 returned error can't find the container with id e4ab74550d5065fc0f9e462ef70761f072a2ea48856618359761cc352b9ba0f8 Jan 29 12:21:55 crc kubenswrapper[4993]: I0129 12:21:55.598461 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" event={"ID":"58863fed-2920-49fd-a1bc-6a5f8a771783","Type":"ContainerStarted","Data":"e4ab74550d5065fc0f9e462ef70761f072a2ea48856618359761cc352b9ba0f8"} Jan 29 12:21:55 crc kubenswrapper[4993]: I0129 12:21:55.975250 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:21:55 crc kubenswrapper[4993]: I0129 12:21:55.976547 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:55 crc kubenswrapper[4993]: I0129 12:21:55.985586 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.124289 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.124538 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.124597 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnq2r\" (UniqueName: \"kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.225518 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnq2r\" (UniqueName: \"kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.225579 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.225614 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.226130 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.226134 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.251333 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnq2r\" (UniqueName: \"kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r\") pod \"certified-operators-jgvgh\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.299725 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:21:56 crc kubenswrapper[4993]: I0129 12:21:56.752931 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:21:57 crc kubenswrapper[4993]: I0129 12:21:57.610353 4993 generic.go:334] "Generic (PLEG): container finished" podID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerID="0d4022c8345f7ae0ca19d598697b729e54f75e40b8516bac3141f12585c94201" exitCode=0 Jan 29 12:21:57 crc kubenswrapper[4993]: I0129 12:21:57.610417 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerDied","Data":"0d4022c8345f7ae0ca19d598697b729e54f75e40b8516bac3141f12585c94201"} Jan 29 12:21:57 crc kubenswrapper[4993]: I0129 12:21:57.611138 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerStarted","Data":"b008cb9f8ddee25f486fd589f237a2a9beffa7e8f36b9bb29ffb1e1212e36dde"} Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.100298 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7wxc2"] Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.101380 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.104053 4993 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b2jjq" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.112937 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7wxc2"] Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.156307 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.156415 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8v4h\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-kube-api-access-z8v4h\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.258030 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.258620 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8v4h\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-kube-api-access-z8v4h\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.288555 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.290624 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8v4h\" (UniqueName: \"kubernetes.io/projected/bfa834cf-9989-403e-b3f8-3ec8b7b62204-kube-api-access-z8v4h\") pod \"cert-manager-cainjector-5545bd876-7wxc2\" (UID: \"bfa834cf-9989-403e-b3f8-3ec8b7b62204\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.435975 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" Jan 29 12:21:58 crc kubenswrapper[4993]: I0129 12:21:58.866813 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7wxc2"] Jan 29 12:21:58 crc kubenswrapper[4993]: W0129 12:21:58.910270 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfa834cf_9989_403e_b3f8_3ec8b7b62204.slice/crio-466a71d19eba5a508c9b640db8544b11429e921b411c14e7718f3fdb83bed26c WatchSource:0}: Error finding container 466a71d19eba5a508c9b640db8544b11429e921b411c14e7718f3fdb83bed26c: Status 404 returned error can't find the container with id 466a71d19eba5a508c9b640db8544b11429e921b411c14e7718f3fdb83bed26c Jan 29 12:21:59 crc kubenswrapper[4993]: I0129 12:21:59.623748 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" event={"ID":"bfa834cf-9989-403e-b3f8-3ec8b7b62204","Type":"ContainerStarted","Data":"466a71d19eba5a508c9b640db8544b11429e921b411c14e7718f3fdb83bed26c"} Jan 29 12:22:03 crc kubenswrapper[4993]: I0129 12:22:03.645523 4993 generic.go:334] "Generic (PLEG): container finished" podID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerID="e6bc59e9acdb782936bb0f5a16a9a5dc1839ed7d43a57a5b8c896664cd01b592" exitCode=0 Jan 29 12:22:03 crc kubenswrapper[4993]: I0129 12:22:03.645603 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerDied","Data":"e6bc59e9acdb782936bb0f5a16a9a5dc1839ed7d43a57a5b8c896664cd01b592"} Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.659227 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerStarted","Data":"d48f8798f0ebcd7e388c636d6e5f54dae741b71d9c8a2a7ae4f2936c6ec2996c"} Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.660711 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" event={"ID":"58863fed-2920-49fd-a1bc-6a5f8a771783","Type":"ContainerStarted","Data":"976ec90e9e53bd36e60506af4f2fef436d03f0b9c5174a0a7dd14a610f219670"} Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.660842 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.661948 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" event={"ID":"bfa834cf-9989-403e-b3f8-3ec8b7b62204","Type":"ContainerStarted","Data":"b0ebf777ace94fafc638236bd673dcb01343d5f9c1b31d056e9a9e565a5a4793"} Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.687545 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jgvgh" podStartSLOduration=3.363381226 podStartE2EDuration="10.687529868s" podCreationTimestamp="2026-01-29 12:21:55 +0000 UTC" firstStartedPulling="2026-01-29 12:21:57.612520901 +0000 UTC m=+1001.627651027" lastFinishedPulling="2026-01-29 12:22:04.936669543 +0000 UTC m=+1008.951799669" observedRunningTime="2026-01-29 12:22:05.685840412 +0000 UTC m=+1009.700970538" watchObservedRunningTime="2026-01-29 12:22:05.687529868 +0000 UTC m=+1009.702659994" Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.706623 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-7wxc2" podStartSLOduration=2.111719657 podStartE2EDuration="7.706607065s" podCreationTimestamp="2026-01-29 12:21:58 +0000 UTC" firstStartedPulling="2026-01-29 12:21:58.913338022 +0000 UTC m=+1002.928468148" lastFinishedPulling="2026-01-29 12:22:04.50822542 +0000 UTC m=+1008.523355556" observedRunningTime="2026-01-29 12:22:05.702069045 +0000 UTC m=+1009.717199171" watchObservedRunningTime="2026-01-29 12:22:05.706607065 +0000 UTC m=+1009.721737191" Jan 29 12:22:05 crc kubenswrapper[4993]: I0129 12:22:05.723453 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" podStartSLOduration=2.799526324 podStartE2EDuration="11.723429033s" podCreationTimestamp="2026-01-29 12:21:54 +0000 UTC" firstStartedPulling="2026-01-29 12:21:55.561878814 +0000 UTC m=+999.577008940" lastFinishedPulling="2026-01-29 12:22:04.485781523 +0000 UTC m=+1008.500911649" observedRunningTime="2026-01-29 12:22:05.718558244 +0000 UTC m=+1009.733688390" watchObservedRunningTime="2026-01-29 12:22:05.723429033 +0000 UTC m=+1009.738559169" Jan 29 12:22:06 crc kubenswrapper[4993]: I0129 12:22:06.300920 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:06 crc kubenswrapper[4993]: I0129 12:22:06.300974 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:07 crc kubenswrapper[4993]: I0129 12:22:07.339590 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jgvgh" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="registry-server" probeResult="failure" output=< Jan 29 12:22:07 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:22:07 crc kubenswrapper[4993]: > Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.000933 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-nr47b"] Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.002021 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.006791 4993 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-n2npm" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.016979 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-nr47b"] Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.096446 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-bound-sa-token\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.096507 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxgx6\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-kube-api-access-lxgx6\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.197124 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxgx6\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-kube-api-access-lxgx6\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.197281 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-bound-sa-token\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.218557 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxgx6\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-kube-api-access-lxgx6\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.226636 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf061abb-955a-43c6-860a-cb439391540e-bound-sa-token\") pod \"cert-manager-545d4d4674-nr47b\" (UID: \"bf061abb-955a-43c6-860a-cb439391540e\") " pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.323224 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-nr47b" Jan 29 12:22:08 crc kubenswrapper[4993]: I0129 12:22:08.744165 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-nr47b"] Jan 29 12:22:09 crc kubenswrapper[4993]: I0129 12:22:09.688324 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-nr47b" event={"ID":"bf061abb-955a-43c6-860a-cb439391540e","Type":"ContainerStarted","Data":"fcbce7399fd8c0efa7941fbbd1401ce31b58226a7f1d33f7994e5f139fb2b4b1"} Jan 29 12:22:09 crc kubenswrapper[4993]: I0129 12:22:09.688675 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-nr47b" event={"ID":"bf061abb-955a-43c6-860a-cb439391540e","Type":"ContainerStarted","Data":"abadbb31e72c1c76c308e29df59471309d37c716a49b8f7ffc9a3af09d1ef3fd"} Jan 29 12:22:09 crc kubenswrapper[4993]: I0129 12:22:09.711504 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-nr47b" podStartSLOduration=2.711483485 podStartE2EDuration="2.711483485s" podCreationTimestamp="2026-01-29 12:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:22:09.707364006 +0000 UTC m=+1013.722494132" watchObservedRunningTime="2026-01-29 12:22:09.711483485 +0000 UTC m=+1013.726613621" Jan 29 12:22:09 crc kubenswrapper[4993]: I0129 12:22:09.910625 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-z9fkp" Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.860940 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.869403 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.874808 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.875700 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.875707 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-snkxz" Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.875964 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 29 12:22:12 crc kubenswrapper[4993]: I0129 12:22:12.964732 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdp8g\" (UniqueName: \"kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g\") pod \"openstack-operator-index-xs2ln\" (UID: \"121e69ea-8e05-44f2-9fb3-aa74bbba6da3\") " pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:13 crc kubenswrapper[4993]: I0129 12:22:13.079762 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdp8g\" (UniqueName: \"kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g\") pod \"openstack-operator-index-xs2ln\" (UID: \"121e69ea-8e05-44f2-9fb3-aa74bbba6da3\") " pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:13 crc kubenswrapper[4993]: I0129 12:22:13.100821 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdp8g\" (UniqueName: \"kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g\") pod \"openstack-operator-index-xs2ln\" (UID: \"121e69ea-8e05-44f2-9fb3-aa74bbba6da3\") " pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:13 crc kubenswrapper[4993]: I0129 12:22:13.194794 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:13 crc kubenswrapper[4993]: I0129 12:22:13.622977 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:13 crc kubenswrapper[4993]: I0129 12:22:13.713934 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xs2ln" event={"ID":"121e69ea-8e05-44f2-9fb3-aa74bbba6da3","Type":"ContainerStarted","Data":"cfd411c999ab484e48e1178f2067485ab67304b0b1511caf0186f166e04d04df"} Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.033599 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.339064 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.379083 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.646773 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7m6nn"] Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.647789 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.654413 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7m6nn"] Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.726303 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlvn9\" (UniqueName: \"kubernetes.io/projected/466ba75a-ab3e-481b-8554-fcf15f73bea6-kube-api-access-qlvn9\") pod \"openstack-operator-index-7m6nn\" (UID: \"466ba75a-ab3e-481b-8554-fcf15f73bea6\") " pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.827438 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlvn9\" (UniqueName: \"kubernetes.io/projected/466ba75a-ab3e-481b-8554-fcf15f73bea6-kube-api-access-qlvn9\") pod \"openstack-operator-index-7m6nn\" (UID: \"466ba75a-ab3e-481b-8554-fcf15f73bea6\") " pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.845897 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlvn9\" (UniqueName: \"kubernetes.io/projected/466ba75a-ab3e-481b-8554-fcf15f73bea6-kube-api-access-qlvn9\") pod \"openstack-operator-index-7m6nn\" (UID: \"466ba75a-ab3e-481b-8554-fcf15f73bea6\") " pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:16 crc kubenswrapper[4993]: I0129 12:22:16.969382 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:17 crc kubenswrapper[4993]: I0129 12:22:17.413756 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7m6nn"] Jan 29 12:22:17 crc kubenswrapper[4993]: I0129 12:22:17.764391 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7m6nn" event={"ID":"466ba75a-ab3e-481b-8554-fcf15f73bea6","Type":"ContainerStarted","Data":"ae2339ea0bc21fd77e7177f1a31edb42f8e9642833c0f75b7d8efa9db6df7818"} Jan 29 12:22:20 crc kubenswrapper[4993]: I0129 12:22:20.436539 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:22:20 crc kubenswrapper[4993]: I0129 12:22:20.438022 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jgvgh" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="registry-server" containerID="cri-o://d48f8798f0ebcd7e388c636d6e5f54dae741b71d9c8a2a7ae4f2936c6ec2996c" gracePeriod=2 Jan 29 12:22:20 crc kubenswrapper[4993]: I0129 12:22:20.798164 4993 generic.go:334] "Generic (PLEG): container finished" podID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerID="d48f8798f0ebcd7e388c636d6e5f54dae741b71d9c8a2a7ae4f2936c6ec2996c" exitCode=0 Jan 29 12:22:20 crc kubenswrapper[4993]: I0129 12:22:20.798226 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerDied","Data":"d48f8798f0ebcd7e388c636d6e5f54dae741b71d9c8a2a7ae4f2936c6ec2996c"} Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.002874 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.091117 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities\") pod \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.091170 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content\") pod \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.091225 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnq2r\" (UniqueName: \"kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r\") pod \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\" (UID: \"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d\") " Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.092792 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities" (OuterVolumeSpecName: "utilities") pod "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" (UID: "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.103586 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r" (OuterVolumeSpecName: "kube-api-access-qnq2r") pod "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" (UID: "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d"). InnerVolumeSpecName "kube-api-access-qnq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.148036 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" (UID: "fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.195971 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.195998 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.196007 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnq2r\" (UniqueName: \"kubernetes.io/projected/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d-kube-api-access-qnq2r\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.814773 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xs2ln" event={"ID":"121e69ea-8e05-44f2-9fb3-aa74bbba6da3","Type":"ContainerStarted","Data":"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a"} Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.814928 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-xs2ln" podUID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" containerName="registry-server" containerID="cri-o://879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a" gracePeriod=2 Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.817904 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7m6nn" event={"ID":"466ba75a-ab3e-481b-8554-fcf15f73bea6","Type":"ContainerStarted","Data":"2eae0cde34d46472bcb1f71bc8cda50f851c5adeb91baff4a3a53c85df9d00a8"} Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.821425 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgvgh" event={"ID":"fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d","Type":"ContainerDied","Data":"b008cb9f8ddee25f486fd589f237a2a9beffa7e8f36b9bb29ffb1e1212e36dde"} Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.821474 4993 scope.go:117] "RemoveContainer" containerID="d48f8798f0ebcd7e388c636d6e5f54dae741b71d9c8a2a7ae4f2936c6ec2996c" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.821882 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgvgh" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.840809 4993 scope.go:117] "RemoveContainer" containerID="e6bc59e9acdb782936bb0f5a16a9a5dc1839ed7d43a57a5b8c896664cd01b592" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.841345 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xs2ln" podStartSLOduration=2.603654571 podStartE2EDuration="9.84131337s" podCreationTimestamp="2026-01-29 12:22:12 +0000 UTC" firstStartedPulling="2026-01-29 12:22:13.633707305 +0000 UTC m=+1017.648837431" lastFinishedPulling="2026-01-29 12:22:20.871366104 +0000 UTC m=+1024.886496230" observedRunningTime="2026-01-29 12:22:21.833994215 +0000 UTC m=+1025.849124371" watchObservedRunningTime="2026-01-29 12:22:21.84131337 +0000 UTC m=+1025.856443496" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.851751 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7m6nn" podStartSLOduration=2.4562095250000002 podStartE2EDuration="5.851728176s" podCreationTimestamp="2026-01-29 12:22:16 +0000 UTC" firstStartedPulling="2026-01-29 12:22:17.444698904 +0000 UTC m=+1021.459829030" lastFinishedPulling="2026-01-29 12:22:20.840217545 +0000 UTC m=+1024.855347681" observedRunningTime="2026-01-29 12:22:21.84697202 +0000 UTC m=+1025.862102146" watchObservedRunningTime="2026-01-29 12:22:21.851728176 +0000 UTC m=+1025.866858302" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.862953 4993 scope.go:117] "RemoveContainer" containerID="0d4022c8345f7ae0ca19d598697b729e54f75e40b8516bac3141f12585c94201" Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.870446 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:22:21 crc kubenswrapper[4993]: I0129 12:22:21.874711 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jgvgh"] Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.189714 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.320062 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdp8g\" (UniqueName: \"kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g\") pod \"121e69ea-8e05-44f2-9fb3-aa74bbba6da3\" (UID: \"121e69ea-8e05-44f2-9fb3-aa74bbba6da3\") " Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.325114 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g" (OuterVolumeSpecName: "kube-api-access-bdp8g") pod "121e69ea-8e05-44f2-9fb3-aa74bbba6da3" (UID: "121e69ea-8e05-44f2-9fb3-aa74bbba6da3"). InnerVolumeSpecName "kube-api-access-bdp8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.422653 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdp8g\" (UniqueName: \"kubernetes.io/projected/121e69ea-8e05-44f2-9fb3-aa74bbba6da3-kube-api-access-bdp8g\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.835596 4993 generic.go:334] "Generic (PLEG): container finished" podID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" containerID="879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a" exitCode=0 Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.835670 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xs2ln" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.835699 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xs2ln" event={"ID":"121e69ea-8e05-44f2-9fb3-aa74bbba6da3","Type":"ContainerDied","Data":"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a"} Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.836087 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xs2ln" event={"ID":"121e69ea-8e05-44f2-9fb3-aa74bbba6da3","Type":"ContainerDied","Data":"cfd411c999ab484e48e1178f2067485ab67304b0b1511caf0186f166e04d04df"} Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.836114 4993 scope.go:117] "RemoveContainer" containerID="879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.869242 4993 scope.go:117] "RemoveContainer" containerID="879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a" Jan 29 12:22:22 crc kubenswrapper[4993]: E0129 12:22:22.871261 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a\": container with ID starting with 879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a not found: ID does not exist" containerID="879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.871313 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a"} err="failed to get container status \"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a\": rpc error: code = NotFound desc = could not find container \"879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a\": container with ID starting with 879b5f13e970f1aa7c2dfb6f03ce76051e5cd8c20e485dc296bbb355f5bd2b1a not found: ID does not exist" Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.872335 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:22 crc kubenswrapper[4993]: I0129 12:22:22.884736 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-xs2ln"] Jan 29 12:22:23 crc kubenswrapper[4993]: I0129 12:22:23.198773 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" path="/var/lib/kubelet/pods/121e69ea-8e05-44f2-9fb3-aa74bbba6da3/volumes" Jan 29 12:22:23 crc kubenswrapper[4993]: I0129 12:22:23.199415 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" path="/var/lib/kubelet/pods/fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d/volumes" Jan 29 12:22:26 crc kubenswrapper[4993]: I0129 12:22:26.970707 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:26 crc kubenswrapper[4993]: I0129 12:22:26.972109 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:27 crc kubenswrapper[4993]: I0129 12:22:27.002966 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:27 crc kubenswrapper[4993]: I0129 12:22:27.899464 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-7m6nn" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.290747 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k"] Jan 29 12:22:34 crc kubenswrapper[4993]: E0129 12:22:34.292164 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="extract-utilities" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.292214 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="extract-utilities" Jan 29 12:22:34 crc kubenswrapper[4993]: E0129 12:22:34.292237 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.292250 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: E0129 12:22:34.292280 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.292294 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: E0129 12:22:34.292316 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="extract-content" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.292328 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="extract-content" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.294256 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="121e69ea-8e05-44f2-9fb3-aa74bbba6da3" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.294305 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa3dc4c2-e83c-436b-a3fb-f23e8a88ec5d" containerName="registry-server" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.295831 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.298102 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-mspj4" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.299671 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k"] Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.386731 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6tpt\" (UniqueName: \"kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.386792 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.386829 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.487648 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6tpt\" (UniqueName: \"kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.487699 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.487731 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.488422 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.488935 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.507569 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6tpt\" (UniqueName: \"kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt\") pod \"c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:34 crc kubenswrapper[4993]: I0129 12:22:34.625462 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:35 crc kubenswrapper[4993]: I0129 12:22:35.021573 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k"] Jan 29 12:22:35 crc kubenswrapper[4993]: W0129 12:22:35.030358 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fdf2bc7_fef0_4ada_b491_f963835bb1cb.slice/crio-2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae WatchSource:0}: Error finding container 2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae: Status 404 returned error can't find the container with id 2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae Jan 29 12:22:35 crc kubenswrapper[4993]: I0129 12:22:35.950745 4993 generic.go:334] "Generic (PLEG): container finished" podID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerID="d8590e91892a80d947bd17bca49b84d0540bc90bdb211baa7f999d2363f23811" exitCode=0 Jan 29 12:22:35 crc kubenswrapper[4993]: I0129 12:22:35.951405 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerDied","Data":"d8590e91892a80d947bd17bca49b84d0540bc90bdb211baa7f999d2363f23811"} Jan 29 12:22:35 crc kubenswrapper[4993]: I0129 12:22:35.953742 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerStarted","Data":"2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae"} Jan 29 12:22:36 crc kubenswrapper[4993]: I0129 12:22:36.964274 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerStarted","Data":"5a81edb49c9f28b67b1045c5be45bc23a27d87056070d38342464a6312c60404"} Jan 29 12:22:37 crc kubenswrapper[4993]: I0129 12:22:37.972715 4993 generic.go:334] "Generic (PLEG): container finished" podID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerID="5a81edb49c9f28b67b1045c5be45bc23a27d87056070d38342464a6312c60404" exitCode=0 Jan 29 12:22:37 crc kubenswrapper[4993]: I0129 12:22:37.972878 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerDied","Data":"5a81edb49c9f28b67b1045c5be45bc23a27d87056070d38342464a6312c60404"} Jan 29 12:22:38 crc kubenswrapper[4993]: I0129 12:22:38.992781 4993 generic.go:334] "Generic (PLEG): container finished" podID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerID="7cd7a590183abfa55ba91ea6a41f71ef4099cb291c80d8f687b5160551d8142a" exitCode=0 Jan 29 12:22:38 crc kubenswrapper[4993]: I0129 12:22:38.992836 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerDied","Data":"7cd7a590183abfa55ba91ea6a41f71ef4099cb291c80d8f687b5160551d8142a"} Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.228464 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.265408 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle\") pod \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.265490 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util\") pod \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.265526 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6tpt\" (UniqueName: \"kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt\") pod \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\" (UID: \"9fdf2bc7-fef0-4ada-b491-f963835bb1cb\") " Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.266565 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle" (OuterVolumeSpecName: "bundle") pod "9fdf2bc7-fef0-4ada-b491-f963835bb1cb" (UID: "9fdf2bc7-fef0-4ada-b491-f963835bb1cb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.272519 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt" (OuterVolumeSpecName: "kube-api-access-l6tpt") pod "9fdf2bc7-fef0-4ada-b491-f963835bb1cb" (UID: "9fdf2bc7-fef0-4ada-b491-f963835bb1cb"). InnerVolumeSpecName "kube-api-access-l6tpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.279242 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util" (OuterVolumeSpecName: "util") pod "9fdf2bc7-fef0-4ada-b491-f963835bb1cb" (UID: "9fdf2bc7-fef0-4ada-b491-f963835bb1cb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.367943 4993 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.368000 4993 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-util\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:40 crc kubenswrapper[4993]: I0129 12:22:40.368032 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6tpt\" (UniqueName: \"kubernetes.io/projected/9fdf2bc7-fef0-4ada-b491-f963835bb1cb-kube-api-access-l6tpt\") on node \"crc\" DevicePath \"\"" Jan 29 12:22:41 crc kubenswrapper[4993]: I0129 12:22:41.006520 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" event={"ID":"9fdf2bc7-fef0-4ada-b491-f963835bb1cb","Type":"ContainerDied","Data":"2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae"} Jan 29 12:22:41 crc kubenswrapper[4993]: I0129 12:22:41.006850 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d6079f2a9488de97e6d90d44b630635b83ce927d4579d1fb9517dd3e5fa22ae" Jan 29 12:22:41 crc kubenswrapper[4993]: I0129 12:22:41.006573 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.478241 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf"] Jan 29 12:22:46 crc kubenswrapper[4993]: E0129 12:22:46.479033 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="extract" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.479047 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="extract" Jan 29 12:22:46 crc kubenswrapper[4993]: E0129 12:22:46.479059 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="pull" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.479065 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="pull" Jan 29 12:22:46 crc kubenswrapper[4993]: E0129 12:22:46.479072 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="util" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.479078 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="util" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.479242 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdf2bc7-fef0-4ada-b491-f963835bb1cb" containerName="extract" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.479710 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.481830 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-ct4gd" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.552227 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscqh\" (UniqueName: \"kubernetes.io/projected/b1d125da-2f26-485e-9010-fd330ee694cd-kube-api-access-pscqh\") pod \"openstack-operator-controller-init-59c8666fb5-c6hrf\" (UID: \"b1d125da-2f26-485e-9010-fd330ee694cd\") " pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.555745 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf"] Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.653666 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscqh\" (UniqueName: \"kubernetes.io/projected/b1d125da-2f26-485e-9010-fd330ee694cd-kube-api-access-pscqh\") pod \"openstack-operator-controller-init-59c8666fb5-c6hrf\" (UID: \"b1d125da-2f26-485e-9010-fd330ee694cd\") " pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.671991 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscqh\" (UniqueName: \"kubernetes.io/projected/b1d125da-2f26-485e-9010-fd330ee694cd-kube-api-access-pscqh\") pod \"openstack-operator-controller-init-59c8666fb5-c6hrf\" (UID: \"b1d125da-2f26-485e-9010-fd330ee694cd\") " pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:46 crc kubenswrapper[4993]: I0129 12:22:46.799727 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:47 crc kubenswrapper[4993]: I0129 12:22:47.007897 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf"] Jan 29 12:22:47 crc kubenswrapper[4993]: I0129 12:22:47.043514 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" event={"ID":"b1d125da-2f26-485e-9010-fd330ee694cd","Type":"ContainerStarted","Data":"ee4890ae2f63b70e3f9cf20fddcfc9b14a054cc362690b3efa0cbd022a92b499"} Jan 29 12:22:54 crc kubenswrapper[4993]: I0129 12:22:54.103078 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" event={"ID":"b1d125da-2f26-485e-9010-fd330ee694cd","Type":"ContainerStarted","Data":"65ff2aa543ec0637064232fe63f63cc8473dfbc24114bc84b16745cd015afe6f"} Jan 29 12:22:54 crc kubenswrapper[4993]: I0129 12:22:54.103739 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:22:54 crc kubenswrapper[4993]: I0129 12:22:54.138873 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" podStartSLOduration=1.925992803 podStartE2EDuration="8.138850794s" podCreationTimestamp="2026-01-29 12:22:46 +0000 UTC" firstStartedPulling="2026-01-29 12:22:47.03181277 +0000 UTC m=+1051.046942896" lastFinishedPulling="2026-01-29 12:22:53.244670751 +0000 UTC m=+1057.259800887" observedRunningTime="2026-01-29 12:22:54.136753714 +0000 UTC m=+1058.151883840" watchObservedRunningTime="2026-01-29 12:22:54.138850794 +0000 UTC m=+1058.153980930" Jan 29 12:23:06 crc kubenswrapper[4993]: I0129 12:23:06.803027 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-59c8666fb5-c6hrf" Jan 29 12:23:18 crc kubenswrapper[4993]: I0129 12:23:18.155576 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:23:18 crc kubenswrapper[4993]: I0129 12:23:18.156178 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.701662 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.703399 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.705417 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-52ztj" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.705997 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.706915 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.710456 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cxpq4" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.715243 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.719896 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.720953 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.725280 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-c5fq4" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.726269 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.744929 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.745830 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.748733 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pn4lz" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.753784 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.766248 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.767417 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.771815 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.776108 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-tngxc" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.792629 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.812489 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.813449 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.820457 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-pwq5x" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.823326 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpgtg\" (UniqueName: \"kubernetes.io/projected/02adb9a3-e33a-4ad6-8563-078804260425-kube-api-access-qpgtg\") pod \"cinder-operator-controller-manager-7595cf584-5zt8x\" (UID: \"02adb9a3-e33a-4ad6-8563-078804260425\") " pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.823553 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk4g2\" (UniqueName: \"kubernetes.io/projected/523bfe53-fa56-448a-9bfe-bc268342e493-kube-api-access-gk4g2\") pod \"barbican-operator-controller-manager-657667746d-hp6j7\" (UID: \"523bfe53-fa56-448a-9bfe-bc268342e493\") " pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.872436 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.920329 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c9n87"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.921245 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.927885 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.927995 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-psppl" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.927890 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pftjm\" (UniqueName: \"kubernetes.io/projected/4ed2c1a4-5753-4802-bbc4-53d89977d7a9-kube-api-access-pftjm\") pod \"horizon-operator-controller-manager-5fb775575f-cp6x6\" (UID: \"4ed2c1a4-5753-4802-bbc4-53d89977d7a9\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.928136 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhmxv\" (UniqueName: \"kubernetes.io/projected/85859ee8-efc0-40b9-a3a0-9ad36c72745f-kube-api-access-fhmxv\") pod \"designate-operator-controller-manager-55d5d5f8ff-846wm\" (UID: \"85859ee8-efc0-40b9-a3a0-9ad36c72745f\") " pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.928226 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk4g2\" (UniqueName: \"kubernetes.io/projected/523bfe53-fa56-448a-9bfe-bc268342e493-kube-api-access-gk4g2\") pod \"barbican-operator-controller-manager-657667746d-hp6j7\" (UID: \"523bfe53-fa56-448a-9bfe-bc268342e493\") " pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.928269 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85fqr\" (UniqueName: \"kubernetes.io/projected/c329c584-dda3-408c-9646-1f7ad46fcff2-kube-api-access-85fqr\") pod \"glance-operator-controller-manager-6db5dbd896-qdnx9\" (UID: \"c329c584-dda3-408c-9646-1f7ad46fcff2\") " pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.928307 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gq2l\" (UniqueName: \"kubernetes.io/projected/3670c965-03c3-40d0-9990-daacad669caa-kube-api-access-8gq2l\") pod \"heat-operator-controller-manager-5499bccc75-9gmx8\" (UID: \"3670c965-03c3-40d0-9990-daacad669caa\") " pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.928369 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpgtg\" (UniqueName: \"kubernetes.io/projected/02adb9a3-e33a-4ad6-8563-078804260425-kube-api-access-qpgtg\") pod \"cinder-operator-controller-manager-7595cf584-5zt8x\" (UID: \"02adb9a3-e33a-4ad6-8563-078804260425\") " pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.935009 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.936175 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.940740 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zh6vs" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.971768 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.977503 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk4g2\" (UniqueName: \"kubernetes.io/projected/523bfe53-fa56-448a-9bfe-bc268342e493-kube-api-access-gk4g2\") pod \"barbican-operator-controller-manager-657667746d-hp6j7\" (UID: \"523bfe53-fa56-448a-9bfe-bc268342e493\") " pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.989245 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c9n87"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.995113 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpgtg\" (UniqueName: \"kubernetes.io/projected/02adb9a3-e33a-4ad6-8563-078804260425-kube-api-access-qpgtg\") pod \"cinder-operator-controller-manager-7595cf584-5zt8x\" (UID: \"02adb9a3-e33a-4ad6-8563-078804260425\") " pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.995229 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs"] Jan 29 12:23:32 crc kubenswrapper[4993]: I0129 12:23:32.996670 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.003621 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-fl2d2" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.003821 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.016069 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.017065 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.017160 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.021694 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.022576 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.022580 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6959m" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.026559 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-hvphh" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.030947 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.031701 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.031928 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.038559 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jjml7" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.040818 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45shv\" (UniqueName: \"kubernetes.io/projected/dcb6de55-31e4-44fc-8e36-d3046fe4d8b2-kube-api-access-45shv\") pod \"ironic-operator-controller-manager-56cb7c4b4c-925f9\" (UID: \"dcb6de55-31e4-44fc-8e36-d3046fe4d8b2\") " pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.040864 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxxx\" (UniqueName: \"kubernetes.io/projected/4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed-kube-api-access-2xxxx\") pod \"neutron-operator-controller-manager-55df775b69-2jnz5\" (UID: \"4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed\") " pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.040937 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pftjm\" (UniqueName: \"kubernetes.io/projected/4ed2c1a4-5753-4802-bbc4-53d89977d7a9-kube-api-access-pftjm\") pod \"horizon-operator-controller-manager-5fb775575f-cp6x6\" (UID: \"4ed2c1a4-5753-4802-bbc4-53d89977d7a9\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.040962 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhmxv\" (UniqueName: \"kubernetes.io/projected/85859ee8-efc0-40b9-a3a0-9ad36c72745f-kube-api-access-fhmxv\") pod \"designate-operator-controller-manager-55d5d5f8ff-846wm\" (UID: \"85859ee8-efc0-40b9-a3a0-9ad36c72745f\") " pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.040990 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgkf2\" (UniqueName: \"kubernetes.io/projected/66e97264-1497-4d2d-968f-e2edc9dc9017-kube-api-access-dgkf2\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041017 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041058 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw4t7\" (UniqueName: \"kubernetes.io/projected/c382415b-73b1-4fe9-a4f0-f828627afb02-kube-api-access-tw4t7\") pod \"keystone-operator-controller-manager-77bb7ffb8c-xlnqs\" (UID: \"c382415b-73b1-4fe9-a4f0-f828627afb02\") " pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041096 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85fqr\" (UniqueName: \"kubernetes.io/projected/c329c584-dda3-408c-9646-1f7ad46fcff2-kube-api-access-85fqr\") pod \"glance-operator-controller-manager-6db5dbd896-qdnx9\" (UID: \"c329c584-dda3-408c-9646-1f7ad46fcff2\") " pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041128 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn8lk\" (UniqueName: \"kubernetes.io/projected/552f4e86-01b8-4c1c-bfdc-e6e7ad197682-kube-api-access-fn8lk\") pod \"manila-operator-controller-manager-6475bdcbc4-8m42n\" (UID: \"552f4e86-01b8-4c1c-bfdc-e6e7ad197682\") " pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041154 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gq2l\" (UniqueName: \"kubernetes.io/projected/3670c965-03c3-40d0-9990-daacad669caa-kube-api-access-8gq2l\") pod \"heat-operator-controller-manager-5499bccc75-9gmx8\" (UID: \"3670c965-03c3-40d0-9990-daacad669caa\") " pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.041180 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r64rg\" (UniqueName: \"kubernetes.io/projected/39304887-7ccc-4b57-95d6-64f886c15e54-kube-api-access-r64rg\") pod \"mariadb-operator-controller-manager-67bf948998-48fq8\" (UID: \"39304887-7ccc-4b57-95d6-64f886c15e54\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.052176 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.055278 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.061826 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.065286 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.066028 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.070690 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wb4g7" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.074253 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.086133 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.087114 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.091007 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhmxv\" (UniqueName: \"kubernetes.io/projected/85859ee8-efc0-40b9-a3a0-9ad36c72745f-kube-api-access-fhmxv\") pod \"designate-operator-controller-manager-55d5d5f8ff-846wm\" (UID: \"85859ee8-efc0-40b9-a3a0-9ad36c72745f\") " pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.095609 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pvgqn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.099233 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.100615 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85fqr\" (UniqueName: \"kubernetes.io/projected/c329c584-dda3-408c-9646-1f7ad46fcff2-kube-api-access-85fqr\") pod \"glance-operator-controller-manager-6db5dbd896-qdnx9\" (UID: \"c329c584-dda3-408c-9646-1f7ad46fcff2\") " pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.100954 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pftjm\" (UniqueName: \"kubernetes.io/projected/4ed2c1a4-5753-4802-bbc4-53d89977d7a9-kube-api-access-pftjm\") pod \"horizon-operator-controller-manager-5fb775575f-cp6x6\" (UID: \"4ed2c1a4-5753-4802-bbc4-53d89977d7a9\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.101596 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gq2l\" (UniqueName: \"kubernetes.io/projected/3670c965-03c3-40d0-9990-daacad669caa-kube-api-access-8gq2l\") pod \"heat-operator-controller-manager-5499bccc75-9gmx8\" (UID: \"3670c965-03c3-40d0-9990-daacad669caa\") " pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.112991 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.113506 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.114097 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.117700 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-hb7w4" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.130522 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.147993 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148570 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45shv\" (UniqueName: \"kubernetes.io/projected/dcb6de55-31e4-44fc-8e36-d3046fe4d8b2-kube-api-access-45shv\") pod \"ironic-operator-controller-manager-56cb7c4b4c-925f9\" (UID: \"dcb6de55-31e4-44fc-8e36-d3046fe4d8b2\") " pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148618 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cg42\" (UniqueName: \"kubernetes.io/projected/aa36f54c-d314-4086-a64e-acec3c80c8bd-kube-api-access-9cg42\") pod \"nova-operator-controller-manager-5ccd5b7f8f-c6gdt\" (UID: \"aa36f54c-d314-4086-a64e-acec3c80c8bd\") " pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148647 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxxx\" (UniqueName: \"kubernetes.io/projected/4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed-kube-api-access-2xxxx\") pod \"neutron-operator-controller-manager-55df775b69-2jnz5\" (UID: \"4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed\") " pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148697 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgkf2\" (UniqueName: \"kubernetes.io/projected/66e97264-1497-4d2d-968f-e2edc9dc9017-kube-api-access-dgkf2\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148723 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148744 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkrms\" (UniqueName: \"kubernetes.io/projected/615c8615-803e-4997-819a-9eecd22cdbee-kube-api-access-pkrms\") pod \"octavia-operator-controller-manager-6b855b4fc4-jrwl8\" (UID: \"615c8615-803e-4997-819a-9eecd22cdbee\") " pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148775 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4t7\" (UniqueName: \"kubernetes.io/projected/c382415b-73b1-4fe9-a4f0-f828627afb02-kube-api-access-tw4t7\") pod \"keystone-operator-controller-manager-77bb7ffb8c-xlnqs\" (UID: \"c382415b-73b1-4fe9-a4f0-f828627afb02\") " pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148802 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-452b8\" (UniqueName: \"kubernetes.io/projected/10455fa4-3adf-4e84-8a37-ae1241d62df6-kube-api-access-452b8\") pod \"ovn-operator-controller-manager-788c46999f-sqf57\" (UID: \"10455fa4-3adf-4e84-8a37-ae1241d62df6\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148826 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn8lk\" (UniqueName: \"kubernetes.io/projected/552f4e86-01b8-4c1c-bfdc-e6e7ad197682-kube-api-access-fn8lk\") pod \"manila-operator-controller-manager-6475bdcbc4-8m42n\" (UID: \"552f4e86-01b8-4c1c-bfdc-e6e7ad197682\") " pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148846 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r64rg\" (UniqueName: \"kubernetes.io/projected/39304887-7ccc-4b57-95d6-64f886c15e54-kube-api-access-r64rg\") pod \"mariadb-operator-controller-manager-67bf948998-48fq8\" (UID: \"39304887-7ccc-4b57-95d6-64f886c15e54\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.148946 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.149451 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.149622 4993 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.149661 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert podName:66e97264-1497-4d2d-968f-e2edc9dc9017 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:33.649645411 +0000 UTC m=+1097.664775527 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert") pod "infra-operator-controller-manager-79955696d6-c9n87" (UID: "66e97264-1497-4d2d-968f-e2edc9dc9017") : secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.149834 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.152548 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-52bgx" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.152700 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.152808 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-llm69" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.152895 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.165671 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.169647 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.185930 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw4t7\" (UniqueName: \"kubernetes.io/projected/c382415b-73b1-4fe9-a4f0-f828627afb02-kube-api-access-tw4t7\") pod \"keystone-operator-controller-manager-77bb7ffb8c-xlnqs\" (UID: \"c382415b-73b1-4fe9-a4f0-f828627afb02\") " pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.188633 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r64rg\" (UniqueName: \"kubernetes.io/projected/39304887-7ccc-4b57-95d6-64f886c15e54-kube-api-access-r64rg\") pod \"mariadb-operator-controller-manager-67bf948998-48fq8\" (UID: \"39304887-7ccc-4b57-95d6-64f886c15e54\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.190150 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.191910 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgkf2\" (UniqueName: \"kubernetes.io/projected/66e97264-1497-4d2d-968f-e2edc9dc9017-kube-api-access-dgkf2\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.192591 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn8lk\" (UniqueName: \"kubernetes.io/projected/552f4e86-01b8-4c1c-bfdc-e6e7ad197682-kube-api-access-fn8lk\") pod \"manila-operator-controller-manager-6475bdcbc4-8m42n\" (UID: \"552f4e86-01b8-4c1c-bfdc-e6e7ad197682\") " pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.201945 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxxx\" (UniqueName: \"kubernetes.io/projected/4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed-kube-api-access-2xxxx\") pod \"neutron-operator-controller-manager-55df775b69-2jnz5\" (UID: \"4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed\") " pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.203372 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45shv\" (UniqueName: \"kubernetes.io/projected/dcb6de55-31e4-44fc-8e36-d3046fe4d8b2-kube-api-access-45shv\") pod \"ironic-operator-controller-manager-56cb7c4b4c-925f9\" (UID: \"dcb6de55-31e4-44fc-8e36-d3046fe4d8b2\") " pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.226487 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.235390 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.236415 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.236510 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.238365 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-dqqwk" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.259013 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkrms\" (UniqueName: \"kubernetes.io/projected/615c8615-803e-4997-819a-9eecd22cdbee-kube-api-access-pkrms\") pod \"octavia-operator-controller-manager-6b855b4fc4-jrwl8\" (UID: \"615c8615-803e-4997-819a-9eecd22cdbee\") " pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.259062 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-452b8\" (UniqueName: \"kubernetes.io/projected/10455fa4-3adf-4e84-8a37-ae1241d62df6-kube-api-access-452b8\") pod \"ovn-operator-controller-manager-788c46999f-sqf57\" (UID: \"10455fa4-3adf-4e84-8a37-ae1241d62df6\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.259117 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cg42\" (UniqueName: \"kubernetes.io/projected/aa36f54c-d314-4086-a64e-acec3c80c8bd-kube-api-access-9cg42\") pod \"nova-operator-controller-manager-5ccd5b7f8f-c6gdt\" (UID: \"aa36f54c-d314-4086-a64e-acec3c80c8bd\") " pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.261228 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.262297 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.265739 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tnjrn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.274283 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.299807 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cg42\" (UniqueName: \"kubernetes.io/projected/aa36f54c-d314-4086-a64e-acec3c80c8bd-kube-api-access-9cg42\") pod \"nova-operator-controller-manager-5ccd5b7f8f-c6gdt\" (UID: \"aa36f54c-d314-4086-a64e-acec3c80c8bd\") " pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.321128 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.339342 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-452b8\" (UniqueName: \"kubernetes.io/projected/10455fa4-3adf-4e84-8a37-ae1241d62df6-kube-api-access-452b8\") pod \"ovn-operator-controller-manager-788c46999f-sqf57\" (UID: \"10455fa4-3adf-4e84-8a37-ae1241d62df6\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.340814 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkrms\" (UniqueName: \"kubernetes.io/projected/615c8615-803e-4997-819a-9eecd22cdbee-kube-api-access-pkrms\") pod \"octavia-operator-controller-manager-6b855b4fc4-jrwl8\" (UID: \"615c8615-803e-4997-819a-9eecd22cdbee\") " pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.361783 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwdp6\" (UniqueName: \"kubernetes.io/projected/30c2b0bb-dbda-49d5-92f6-567703290c05-kube-api-access-rwdp6\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.361980 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p95c4\" (UniqueName: \"kubernetes.io/projected/0a0fde08-a591-4fb8-a872-c0a4cd7e2758-kube-api-access-p95c4\") pod \"swift-operator-controller-manager-6f7455757b-ls4r8\" (UID: \"0a0fde08-a591-4fb8-a872-c0a4cd7e2758\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.364462 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.364510 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkvjp\" (UniqueName: \"kubernetes.io/projected/66cb8608-8dd0-4304-9245-8c4468d7c6eb-kube-api-access-tkvjp\") pod \"placement-operator-controller-manager-5b964cf4cd-2xcgn\" (UID: \"66cb8608-8dd0-4304-9245-8c4468d7c6eb\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.390532 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.391048 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.434906 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.464251 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.465300 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.466662 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-997f9\" (UniqueName: \"kubernetes.io/projected/c3df3d31-0166-4aaa-b6a8-98865bc4d99c-kube-api-access-997f9\") pod \"telemetry-operator-controller-manager-c95fd9dc5-fqltq\" (UID: \"c3df3d31-0166-4aaa-b6a8-98865bc4d99c\") " pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.466700 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.466735 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkvjp\" (UniqueName: \"kubernetes.io/projected/66cb8608-8dd0-4304-9245-8c4468d7c6eb-kube-api-access-tkvjp\") pod \"placement-operator-controller-manager-5b964cf4cd-2xcgn\" (UID: \"66cb8608-8dd0-4304-9245-8c4468d7c6eb\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.466895 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwdp6\" (UniqueName: \"kubernetes.io/projected/30c2b0bb-dbda-49d5-92f6-567703290c05-kube-api-access-rwdp6\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.466966 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p95c4\" (UniqueName: \"kubernetes.io/projected/0a0fde08-a591-4fb8-a872-c0a4cd7e2758-kube-api-access-p95c4\") pod \"swift-operator-controller-manager-6f7455757b-ls4r8\" (UID: \"0a0fde08-a591-4fb8-a872-c0a4cd7e2758\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.466991 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.467033 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:33.967018986 +0000 UTC m=+1097.982149202 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.467593 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qhgsf" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.477761 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.503238 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.504908 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkvjp\" (UniqueName: \"kubernetes.io/projected/66cb8608-8dd0-4304-9245-8c4468d7c6eb-kube-api-access-tkvjp\") pod \"placement-operator-controller-manager-5b964cf4cd-2xcgn\" (UID: \"66cb8608-8dd0-4304-9245-8c4468d7c6eb\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.535171 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p95c4\" (UniqueName: \"kubernetes.io/projected/0a0fde08-a591-4fb8-a872-c0a4cd7e2758-kube-api-access-p95c4\") pod \"swift-operator-controller-manager-6f7455757b-ls4r8\" (UID: \"0a0fde08-a591-4fb8-a872-c0a4cd7e2758\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.538641 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwdp6\" (UniqueName: \"kubernetes.io/projected/30c2b0bb-dbda-49d5-92f6-567703290c05-kube-api-access-rwdp6\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.548629 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.565425 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.568252 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qtlw\" (UniqueName: \"kubernetes.io/projected/be15d7ad-6e9d-40fe-845e-626f4b1dc766-kube-api-access-9qtlw\") pod \"test-operator-controller-manager-56f8bfcd9f-m9xms\" (UID: \"be15d7ad-6e9d-40fe-845e-626f4b1dc766\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.568310 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-997f9\" (UniqueName: \"kubernetes.io/projected/c3df3d31-0166-4aaa-b6a8-98865bc4d99c-kube-api-access-997f9\") pod \"telemetry-operator-controller-manager-c95fd9dc5-fqltq\" (UID: \"c3df3d31-0166-4aaa-b6a8-98865bc4d99c\") " pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.585088 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.598340 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.599423 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.616168 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.617660 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vpw4b" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.636688 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.637058 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.637586 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.640161 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.648388 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-997f9\" (UniqueName: \"kubernetes.io/projected/c3df3d31-0166-4aaa-b6a8-98865bc4d99c-kube-api-access-997f9\") pod \"telemetry-operator-controller-manager-c95fd9dc5-fqltq\" (UID: \"c3df3d31-0166-4aaa-b6a8-98865bc4d99c\") " pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.652922 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.653157 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7wslb" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.653602 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.657767 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.659655 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.672211 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-czlv2" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.672650 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.676885 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.702794 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.703076 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qtlw\" (UniqueName: \"kubernetes.io/projected/be15d7ad-6e9d-40fe-845e-626f4b1dc766-kube-api-access-9qtlw\") pod \"test-operator-controller-manager-56f8bfcd9f-m9xms\" (UID: \"be15d7ad-6e9d-40fe-845e-626f4b1dc766\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.703273 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqr7l\" (UniqueName: \"kubernetes.io/projected/181f3d8c-9fd7-4df6-806a-8d04ac0ff597-kube-api-access-lqr7l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pxqq5\" (UID: \"181f3d8c-9fd7-4df6-806a-8d04ac0ff597\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.703707 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.703758 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svx2l\" (UniqueName: \"kubernetes.io/projected/6c7a911f-47c8-4fab-adb0-7d179c090b29-kube-api-access-svx2l\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.705800 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.706639 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.707565 4993 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.707626 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert podName:66e97264-1497-4d2d-968f-e2edc9dc9017 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:34.707607699 +0000 UTC m=+1098.722737825 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert") pod "infra-operator-controller-manager-79955696d6-c9n87" (UID: "66e97264-1497-4d2d-968f-e2edc9dc9017") : secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.748518 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.761995 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qtlw\" (UniqueName: \"kubernetes.io/projected/be15d7ad-6e9d-40fe-845e-626f4b1dc766-kube-api-access-9qtlw\") pod \"test-operator-controller-manager-56f8bfcd9f-m9xms\" (UID: \"be15d7ad-6e9d-40fe-845e-626f4b1dc766\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.794701 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.810939 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqr7l\" (UniqueName: \"kubernetes.io/projected/181f3d8c-9fd7-4df6-806a-8d04ac0ff597-kube-api-access-lqr7l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pxqq5\" (UID: \"181f3d8c-9fd7-4df6-806a-8d04ac0ff597\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.811014 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.811040 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7qgm\" (UniqueName: \"kubernetes.io/projected/da5d8a04-0e63-4b7a-b1b7-460a97473fce-kube-api-access-m7qgm\") pod \"watcher-operator-controller-manager-56b5dc77fd-v57np\" (UID: \"da5d8a04-0e63-4b7a-b1b7-460a97473fce\") " pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.811062 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svx2l\" (UniqueName: \"kubernetes.io/projected/6c7a911f-47c8-4fab-adb0-7d179c090b29-kube-api-access-svx2l\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.811082 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.811257 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.811308 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:34.311292753 +0000 UTC m=+1098.326422889 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.812058 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: E0129 12:23:33.812129 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:34.312112356 +0000 UTC m=+1098.327242492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.839974 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x"] Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.859099 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svx2l\" (UniqueName: \"kubernetes.io/projected/6c7a911f-47c8-4fab-adb0-7d179c090b29-kube-api-access-svx2l\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.859961 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqr7l\" (UniqueName: \"kubernetes.io/projected/181f3d8c-9fd7-4df6-806a-8d04ac0ff597-kube-api-access-lqr7l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-pxqq5\" (UID: \"181f3d8c-9fd7-4df6-806a-8d04ac0ff597\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.912289 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7qgm\" (UniqueName: \"kubernetes.io/projected/da5d8a04-0e63-4b7a-b1b7-460a97473fce-kube-api-access-m7qgm\") pod \"watcher-operator-controller-manager-56b5dc77fd-v57np\" (UID: \"da5d8a04-0e63-4b7a-b1b7-460a97473fce\") " pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.942900 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7qgm\" (UniqueName: \"kubernetes.io/projected/da5d8a04-0e63-4b7a-b1b7-460a97473fce-kube-api-access-m7qgm\") pod \"watcher-operator-controller-manager-56b5dc77fd-v57np\" (UID: \"da5d8a04-0e63-4b7a-b1b7-460a97473fce\") " pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:33 crc kubenswrapper[4993]: I0129 12:23:33.972158 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:33.990052 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.013618 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.014579 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.014623 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:35.014608526 +0000 UTC m=+1099.029738652 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.318886 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.319155 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.319408 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.319464 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:35.319446812 +0000 UTC m=+1099.334576938 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.319848 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.319881 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:35.319871514 +0000 UTC m=+1099.335001640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.336381 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.350577 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.424905 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" event={"ID":"4ed2c1a4-5753-4802-bbc4-53d89977d7a9","Type":"ContainerStarted","Data":"a60b45f90e2b5405c88dbbc4844a3ad51eeda35fda992cf8375f8b60fb0ecd51"} Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.426232 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" event={"ID":"dcb6de55-31e4-44fc-8e36-d3046fe4d8b2","Type":"ContainerStarted","Data":"8296250e662e1a1d39e7a71fe7b98d77926a589a5170750057f32598152c241c"} Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.427331 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" event={"ID":"02adb9a3-e33a-4ad6-8563-078804260425","Type":"ContainerStarted","Data":"ccc67cd52c22e1bd9d9a45f35ed70bbb6bea49692e288ac4c42889b6131ede7e"} Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.428618 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" event={"ID":"523bfe53-fa56-448a-9bfe-bc268342e493","Type":"ContainerStarted","Data":"d634a0f7d9a0a35efa884418940f5b8a8faf916bb78ec6bff68197f2db55dbfd"} Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.725090 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.725290 4993 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.725349 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert podName:66e97264-1497-4d2d-968f-e2edc9dc9017 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:36.725329065 +0000 UTC m=+1100.740459191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert") pod "infra-operator-controller-manager-79955696d6-c9n87" (UID: "66e97264-1497-4d2d-968f-e2edc9dc9017") : secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.745612 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8"] Jan 29 12:23:34 crc kubenswrapper[4993]: W0129 12:23:34.752232 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b6f17b9_04e0_44b0_bd5d_7cb91cc9efed.slice/crio-20f467eca2d3aa125522e1db14f2c55e7f6888acb26c4e4adfa18828a535b919 WatchSource:0}: Error finding container 20f467eca2d3aa125522e1db14f2c55e7f6888acb26c4e4adfa18828a535b919: Status 404 returned error can't find the container with id 20f467eca2d3aa125522e1db14f2c55e7f6888acb26c4e4adfa18828a535b919 Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.786415 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.802735 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.831720 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.844297 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm"] Jan 29 12:23:34 crc kubenswrapper[4993]: W0129 12:23:34.846611 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod552f4e86_01b8_4c1c_bfdc_e6e7ad197682.slice/crio-8c9166a37d773fb7186f79bd695bf8fd5536a0eda32146f50741266ce04c1b76 WatchSource:0}: Error finding container 8c9166a37d773fb7186f79bd695bf8fd5536a0eda32146f50741266ce04c1b76: Status 404 returned error can't find the container with id 8c9166a37d773fb7186f79bd695bf8fd5536a0eda32146f50741266ce04c1b76 Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.862891 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.876867 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.883342 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.890044 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.893823 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8"] Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.927967 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lqr7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-pxqq5_openstack-operators(181f3d8c-9fd7-4df6-806a-8d04ac0ff597): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.929224 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podUID="181f3d8c-9fd7-4df6-806a-8d04ac0ff597" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.932827 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5"] Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.934606 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9cg42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5ccd5b7f8f-c6gdt_openstack-operators(aa36f54c-d314-4086-a64e-acec3c80c8bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.936150 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podUID="aa36f54c-d314-4086-a64e-acec3c80c8bd" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.941865 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt"] Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.945311 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkvjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b964cf4cd-2xcgn_openstack-operators(66cb8608-8dd0-4304-9245-8c4468d7c6eb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.947049 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podUID="66cb8608-8dd0-4304-9245-8c4468d7c6eb" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.948486 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-997f9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-c95fd9dc5-fqltq_openstack-operators(c3df3d31-0166-4aaa-b6a8-98865bc4d99c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.949645 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57"] Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.949912 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podUID="c3df3d31-0166-4aaa-b6a8-98865bc4d99c" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.953635 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p95c4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6f7455757b-ls4r8_openstack-operators(0a0fde08-a591-4fb8-a872-c0a4cd7e2758): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.953785 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-452b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-788c46999f-sqf57_openstack-operators(10455fa4-3adf-4e84-8a37-ae1241d62df6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.955879 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" podUID="10455fa4-3adf-4e84-8a37-ae1241d62df6" Jan 29 12:23:34 crc kubenswrapper[4993]: E0129 12:23:34.956017 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podUID="0a0fde08-a591-4fb8-a872-c0a4cd7e2758" Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.960170 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.968568 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq"] Jan 29 12:23:34 crc kubenswrapper[4993]: I0129 12:23:34.993884 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8"] Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.030549 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.030722 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.030816 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:37.03079335 +0000 UTC m=+1101.045923476 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.334884 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.335253 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.335059 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.335549 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:37.335521611 +0000 UTC m=+1101.350651747 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.335474 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.335650 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:37.335607324 +0000 UTC m=+1101.350737520 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.436470 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" event={"ID":"aa36f54c-d314-4086-a64e-acec3c80c8bd","Type":"ContainerStarted","Data":"8197f6bdcf38225bb4349700ab792297181aa613391ff459470b93ce5df2e23b"} Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.439368 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podUID="aa36f54c-d314-4086-a64e-acec3c80c8bd" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.441057 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" event={"ID":"0a0fde08-a591-4fb8-a872-c0a4cd7e2758","Type":"ContainerStarted","Data":"3f9c6cc293cdbe06251eb7d0b3a093e91a783dd6632af055bcf271790baf3177"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.442782 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" event={"ID":"10455fa4-3adf-4e84-8a37-ae1241d62df6","Type":"ContainerStarted","Data":"3ee9768ca8a8586cd58aa3e9805c547a3a3925b54478219d0b813d5ad62158f3"} Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.443013 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podUID="0a0fde08-a591-4fb8-a872-c0a4cd7e2758" Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.443879 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" podUID="10455fa4-3adf-4e84-8a37-ae1241d62df6" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.444130 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" event={"ID":"552f4e86-01b8-4c1c-bfdc-e6e7ad197682","Type":"ContainerStarted","Data":"8c9166a37d773fb7186f79bd695bf8fd5536a0eda32146f50741266ce04c1b76"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.445393 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" event={"ID":"39304887-7ccc-4b57-95d6-64f886c15e54","Type":"ContainerStarted","Data":"77f847432b190078a575e66b3d2f8271414737517efba34a548391c53ccefdbb"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.446401 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" event={"ID":"c382415b-73b1-4fe9-a4f0-f828627afb02","Type":"ContainerStarted","Data":"6048b5ea5a3bd26dfb06a721fd1fbab2dc7d3d6ddef7eb6f60052b6c7f467cf2"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.450094 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" event={"ID":"181f3d8c-9fd7-4df6-806a-8d04ac0ff597","Type":"ContainerStarted","Data":"fa03eb3af062dce32f3001c6e416f7a9d661269449d2a0355c3fb3792e06849c"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.455270 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" event={"ID":"da5d8a04-0e63-4b7a-b1b7-460a97473fce","Type":"ContainerStarted","Data":"8a8ab85acc21771ca67f0306dd2668f84cc00e3095835b5373d43e11f18cd9a6"} Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.457666 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podUID="181f3d8c-9fd7-4df6-806a-8d04ac0ff597" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.458791 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" event={"ID":"66cb8608-8dd0-4304-9245-8c4468d7c6eb","Type":"ContainerStarted","Data":"92c8e49accdfc1929b7d3f9fa42025219bd24dc8cd6da7b9196e887859dca6b8"} Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.461374 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podUID="66cb8608-8dd0-4304-9245-8c4468d7c6eb" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.463070 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" event={"ID":"c329c584-dda3-408c-9646-1f7ad46fcff2","Type":"ContainerStarted","Data":"3704134d601796b68de0dcad6d06f9f29d16f348d9d632446c8a10bec10af64a"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.467997 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" event={"ID":"be15d7ad-6e9d-40fe-845e-626f4b1dc766","Type":"ContainerStarted","Data":"fec5e9b90a48ca7e185a7978eaec576de645393346da9797e3739eb9995d3d6e"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.478317 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" event={"ID":"c3df3d31-0166-4aaa-b6a8-98865bc4d99c","Type":"ContainerStarted","Data":"39eb681109cbe4b8bac0c126ee92a9a47d495041011c85f9762b261c16d7b669"} Jan 29 12:23:35 crc kubenswrapper[4993]: E0129 12:23:35.479754 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podUID="c3df3d31-0166-4aaa-b6a8-98865bc4d99c" Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.483861 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" event={"ID":"615c8615-803e-4997-819a-9eecd22cdbee","Type":"ContainerStarted","Data":"649f690b4384028f9aebe751d6950b3e8714749e9b5bde39ecabdb963208cd0e"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.485963 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" event={"ID":"4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed","Type":"ContainerStarted","Data":"20f467eca2d3aa125522e1db14f2c55e7f6888acb26c4e4adfa18828a535b919"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.487060 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" event={"ID":"85859ee8-efc0-40b9-a3a0-9ad36c72745f","Type":"ContainerStarted","Data":"f2cd6d2c0b6dcf959f98ceb5af69a33d94f984a03a48da87f0aaf227ff5e8eef"} Jan 29 12:23:35 crc kubenswrapper[4993]: I0129 12:23:35.488166 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" event={"ID":"3670c965-03c3-40d0-9990-daacad669caa","Type":"ContainerStarted","Data":"016e19b1bc979bd84ca158715d1d46e700a38796e8e8fafbe06b8de523018121"} Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.501491 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podUID="0a0fde08-a591-4fb8-a872-c0a4cd7e2758" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.501915 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podUID="181f3d8c-9fd7-4df6-806a-8d04ac0ff597" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.501971 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" podUID="10455fa4-3adf-4e84-8a37-ae1241d62df6" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.503484 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podUID="66cb8608-8dd0-4304-9245-8c4468d7c6eb" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.503569 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podUID="c3df3d31-0166-4aaa-b6a8-98865bc4d99c" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.503626 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podUID="aa36f54c-d314-4086-a64e-acec3c80c8bd" Jan 29 12:23:36 crc kubenswrapper[4993]: I0129 12:23:36.781007 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.781149 4993 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:36 crc kubenswrapper[4993]: E0129 12:23:36.781300 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert podName:66e97264-1497-4d2d-968f-e2edc9dc9017 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:40.781275408 +0000 UTC m=+1104.796405534 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert") pod "infra-operator-controller-manager-79955696d6-c9n87" (UID: "66e97264-1497-4d2d-968f-e2edc9dc9017") : secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: I0129 12:23:37.089664 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.090108 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.090155 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:41.090141209 +0000 UTC m=+1105.105271335 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: I0129 12:23:37.396673 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:37 crc kubenswrapper[4993]: I0129 12:23:37.396728 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.396923 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.397008 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:41.396984393 +0000 UTC m=+1105.412114569 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.397508 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:37 crc kubenswrapper[4993]: E0129 12:23:37.397552 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:41.397539688 +0000 UTC m=+1105.412669894 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:40 crc kubenswrapper[4993]: I0129 12:23:40.864043 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:40 crc kubenswrapper[4993]: E0129 12:23:40.864367 4993 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:40 crc kubenswrapper[4993]: E0129 12:23:40.864482 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert podName:66e97264-1497-4d2d-968f-e2edc9dc9017 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:48.864462933 +0000 UTC m=+1112.879593059 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert") pod "infra-operator-controller-manager-79955696d6-c9n87" (UID: "66e97264-1497-4d2d-968f-e2edc9dc9017") : secret "infra-operator-webhook-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: I0129 12:23:41.168162 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.168325 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.168399 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:49.168381888 +0000 UTC m=+1113.183512014 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: I0129 12:23:41.472623 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:41 crc kubenswrapper[4993]: I0129 12:23:41.472678 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.472785 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.472802 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.472840 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:49.472827059 +0000 UTC m=+1113.487957185 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:41 crc kubenswrapper[4993]: E0129 12:23:41.472853 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:23:49.472848199 +0000 UTC m=+1113.487978325 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.596474 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" event={"ID":"39304887-7ccc-4b57-95d6-64f886c15e54","Type":"ContainerStarted","Data":"5b7cf34ce6d5d3c83cb5335952d90b7ee5bc866a1f363f16cd08ce69ac0ab182"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.597072 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.597978 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" event={"ID":"85859ee8-efc0-40b9-a3a0-9ad36c72745f","Type":"ContainerStarted","Data":"03aa7ad3d56bdaf0e79d92e36a3e81366f8945681dd3a80f67fed4b1e8f10e2c"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.598397 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.599562 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" event={"ID":"523bfe53-fa56-448a-9bfe-bc268342e493","Type":"ContainerStarted","Data":"6f4eefb2dcc8e3b35c260446a46d126d58f6773ac59ae6e189cf14c1cc52d879"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.599624 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.601804 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" event={"ID":"c382415b-73b1-4fe9-a4f0-f828627afb02","Type":"ContainerStarted","Data":"5fb2615e83918ccb9951c54be6dc2477afaa6980632a3cb539912339a13c1bcc"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.601936 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.603388 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" event={"ID":"4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed","Type":"ContainerStarted","Data":"dc1c679095e10f13aa15c6e59a9be17a791a23df62a0877afbbddcc5d912ed4a"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.603482 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.604824 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" event={"ID":"dcb6de55-31e4-44fc-8e36-d3046fe4d8b2","Type":"ContainerStarted","Data":"7eaca2e36676ea739c92068ed19e83c70e710bdf2efbdb5f69eb43be007cfab7"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.604934 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.606366 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" event={"ID":"02adb9a3-e33a-4ad6-8563-078804260425","Type":"ContainerStarted","Data":"d51f61974b2878fc36f658e7a55fbe62dcc73cfe5db1b6b742c419cdb58a0250"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.606507 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.607906 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" event={"ID":"c329c584-dda3-408c-9646-1f7ad46fcff2","Type":"ContainerStarted","Data":"00d0a94bd686e7bdc101b62be1e2ecedb17cc316925cf1aa0501096a0a6dd776"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.608251 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.610774 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" event={"ID":"3670c965-03c3-40d0-9990-daacad669caa","Type":"ContainerStarted","Data":"b59135944094f4de1ce4d1cd3a1413e7275b8da7ad8093e23d3bc013a4793d5e"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.611141 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.612987 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" event={"ID":"be15d7ad-6e9d-40fe-845e-626f4b1dc766","Type":"ContainerStarted","Data":"e8816d65349dfd0d22582104bad5780129322295205b5b12e4aaddacbcea9138"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.613359 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.614676 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" event={"ID":"615c8615-803e-4997-819a-9eecd22cdbee","Type":"ContainerStarted","Data":"5dbd542bf5d774c6aec1b48eb44081bf496dd763d0b881821acffafc1da0c3b5"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.622575 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.645850 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" event={"ID":"4ed2c1a4-5753-4802-bbc4-53d89977d7a9","Type":"ContainerStarted","Data":"21f6e67183fc306bb2ef85e80b0b930a50483c14de33e444d987d430dd33b15a"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.647016 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.655511 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" event={"ID":"552f4e86-01b8-4c1c-bfdc-e6e7ad197682","Type":"ContainerStarted","Data":"a3f0b8c7088424fb478876f751dad43f2d6e7931e52ac70833251fa337902545"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.656727 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.661801 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" event={"ID":"da5d8a04-0e63-4b7a-b1b7-460a97473fce","Type":"ContainerStarted","Data":"c9d73d6f33c3bfbeea26382177394c801a8fbc69338589ea011655e85eba95dd"} Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.662148 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" podStartSLOduration=3.993468576 podStartE2EDuration="15.662124584s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.752717842 +0000 UTC m=+1098.767847968" lastFinishedPulling="2026-01-29 12:23:46.42137385 +0000 UTC m=+1110.436503976" observedRunningTime="2026-01-29 12:23:47.636874078 +0000 UTC m=+1111.652004204" watchObservedRunningTime="2026-01-29 12:23:47.662124584 +0000 UTC m=+1111.677254710" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.662980 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.680318 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" podStartSLOduration=3.279400504 podStartE2EDuration="15.680303417s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.022278776 +0000 UTC m=+1098.037408902" lastFinishedPulling="2026-01-29 12:23:46.423181689 +0000 UTC m=+1110.438311815" observedRunningTime="2026-01-29 12:23:47.678361725 +0000 UTC m=+1111.693491851" watchObservedRunningTime="2026-01-29 12:23:47.680303417 +0000 UTC m=+1111.695433543" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.704889 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" podStartSLOduration=3.65768509 podStartE2EDuration="15.704870545s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.362986721 +0000 UTC m=+1098.378116847" lastFinishedPulling="2026-01-29 12:23:46.410172186 +0000 UTC m=+1110.425302302" observedRunningTime="2026-01-29 12:23:47.698993815 +0000 UTC m=+1111.714123941" watchObservedRunningTime="2026-01-29 12:23:47.704870545 +0000 UTC m=+1111.720000681" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.744441 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" podStartSLOduration=3.315340668 podStartE2EDuration="15.744421459s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:33.98166352 +0000 UTC m=+1097.996793646" lastFinishedPulling="2026-01-29 12:23:46.410744291 +0000 UTC m=+1110.425874437" observedRunningTime="2026-01-29 12:23:47.743945337 +0000 UTC m=+1111.759075473" watchObservedRunningTime="2026-01-29 12:23:47.744421459 +0000 UTC m=+1111.759551585" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.804498 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" podStartSLOduration=4.216208738 podStartE2EDuration="15.804477941s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.832223353 +0000 UTC m=+1098.847353479" lastFinishedPulling="2026-01-29 12:23:46.420492556 +0000 UTC m=+1110.435622682" observedRunningTime="2026-01-29 12:23:47.776044898 +0000 UTC m=+1111.791175024" watchObservedRunningTime="2026-01-29 12:23:47.804477941 +0000 UTC m=+1111.819608077" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.804863 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" podStartSLOduration=4.122387506 podStartE2EDuration="15.804857241s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.771533191 +0000 UTC m=+1098.786663317" lastFinishedPulling="2026-01-29 12:23:46.454002926 +0000 UTC m=+1110.469133052" observedRunningTime="2026-01-29 12:23:47.790551442 +0000 UTC m=+1111.805681568" watchObservedRunningTime="2026-01-29 12:23:47.804857241 +0000 UTC m=+1111.819987367" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.831532 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" podStartSLOduration=3.248001968 podStartE2EDuration="14.831516505s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.83701373 +0000 UTC m=+1098.852143856" lastFinishedPulling="2026-01-29 12:23:46.420528277 +0000 UTC m=+1110.435658393" observedRunningTime="2026-01-29 12:23:47.82987034 +0000 UTC m=+1111.845000466" watchObservedRunningTime="2026-01-29 12:23:47.831516505 +0000 UTC m=+1111.846646631" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.850316 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" podStartSLOduration=4.273492975 podStartE2EDuration="15.850301065s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.846565354 +0000 UTC m=+1098.861695480" lastFinishedPulling="2026-01-29 12:23:46.423373444 +0000 UTC m=+1110.438503570" observedRunningTime="2026-01-29 12:23:47.846230565 +0000 UTC m=+1111.861360681" watchObservedRunningTime="2026-01-29 12:23:47.850301065 +0000 UTC m=+1111.865431191" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.910961 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" podStartSLOduration=4.294200576 podStartE2EDuration="15.910938983s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.792591126 +0000 UTC m=+1098.807721262" lastFinishedPulling="2026-01-29 12:23:46.409329553 +0000 UTC m=+1110.424459669" observedRunningTime="2026-01-29 12:23:47.87256957 +0000 UTC m=+1111.887699696" watchObservedRunningTime="2026-01-29 12:23:47.910938983 +0000 UTC m=+1111.926069109" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.920379 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" podStartSLOduration=4.271775686 podStartE2EDuration="15.920361009s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.771961634 +0000 UTC m=+1098.787091760" lastFinishedPulling="2026-01-29 12:23:46.420546957 +0000 UTC m=+1110.435677083" observedRunningTime="2026-01-29 12:23:47.910725407 +0000 UTC m=+1111.925855543" watchObservedRunningTime="2026-01-29 12:23:47.920361009 +0000 UTC m=+1111.935491125" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.937616 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" podStartSLOduration=4.257841338 podStartE2EDuration="15.937601447s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.760137464 +0000 UTC m=+1098.775267590" lastFinishedPulling="2026-01-29 12:23:46.439897573 +0000 UTC m=+1110.455027699" observedRunningTime="2026-01-29 12:23:47.934116722 +0000 UTC m=+1111.949246848" watchObservedRunningTime="2026-01-29 12:23:47.937601447 +0000 UTC m=+1111.952731573" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.955668 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" podStartSLOduration=3.350611555 podStartE2EDuration="14.955653727s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.803999323 +0000 UTC m=+1098.819129449" lastFinishedPulling="2026-01-29 12:23:46.409041495 +0000 UTC m=+1110.424171621" observedRunningTime="2026-01-29 12:23:47.953266312 +0000 UTC m=+1111.968396438" watchObservedRunningTime="2026-01-29 12:23:47.955653727 +0000 UTC m=+1111.970783853" Jan 29 12:23:47 crc kubenswrapper[4993]: I0129 12:23:47.994697 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" podStartSLOduration=4.420856574 podStartE2EDuration="15.994682218s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.836757363 +0000 UTC m=+1098.851887499" lastFinishedPulling="2026-01-29 12:23:46.410583017 +0000 UTC m=+1110.425713143" observedRunningTime="2026-01-29 12:23:47.993419183 +0000 UTC m=+1112.008549309" watchObservedRunningTime="2026-01-29 12:23:47.994682218 +0000 UTC m=+1112.009812344" Jan 29 12:23:48 crc kubenswrapper[4993]: I0129 12:23:48.014757 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" podStartSLOduration=3.970299036 podStartE2EDuration="16.014737122s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.378871246 +0000 UTC m=+1098.394001362" lastFinishedPulling="2026-01-29 12:23:46.423309322 +0000 UTC m=+1110.438439448" observedRunningTime="2026-01-29 12:23:48.010127667 +0000 UTC m=+1112.025257793" watchObservedRunningTime="2026-01-29 12:23:48.014737122 +0000 UTC m=+1112.029867248" Jan 29 12:23:48 crc kubenswrapper[4993]: I0129 12:23:48.156103 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:23:48 crc kubenswrapper[4993]: I0129 12:23:48.156386 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:23:48 crc kubenswrapper[4993]: I0129 12:23:48.893341 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:48 crc kubenswrapper[4993]: I0129 12:23:48.901588 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/66e97264-1497-4d2d-968f-e2edc9dc9017-cert\") pod \"infra-operator-controller-manager-79955696d6-c9n87\" (UID: \"66e97264-1497-4d2d-968f-e2edc9dc9017\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:49 crc kubenswrapper[4993]: I0129 12:23:49.147142 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:23:49 crc kubenswrapper[4993]: I0129 12:23:49.198633 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.198917 4993 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.198999 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert podName:30c2b0bb-dbda-49d5-92f6-567703290c05 nodeName:}" failed. No retries permitted until 2026-01-29 12:24:05.198980361 +0000 UTC m=+1129.214110487 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" (UID: "30c2b0bb-dbda-49d5-92f6-567703290c05") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: I0129 12:23:49.504019 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.504642 4993 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.504728 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:24:05.504703565 +0000 UTC m=+1129.519833761 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "metrics-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: I0129 12:23:49.504653 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.505374 4993 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: E0129 12:23:49.505471 4993 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs podName:6c7a911f-47c8-4fab-adb0-7d179c090b29 nodeName:}" failed. No retries permitted until 2026-01-29 12:24:05.505451086 +0000 UTC m=+1129.520581282 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs") pod "openstack-operator-controller-manager-65dc8f5954-7dnzd" (UID: "6c7a911f-47c8-4fab-adb0-7d179c090b29") : secret "webhook-server-cert" not found Jan 29 12:23:49 crc kubenswrapper[4993]: I0129 12:23:49.878909 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c9n87"] Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.035311 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-657667746d-hp6j7" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.055566 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7595cf584-5zt8x" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.119875 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5499bccc75-9gmx8" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.173312 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cp6x6" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.198927 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-48fq8" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.229754 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-55df775b69-2jnz5" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.277462 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-56cb7c4b4c-925f9" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.394264 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-55d5d5f8ff-846wm" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.394564 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6db5dbd896-qdnx9" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.453309 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-77bb7ffb8c-xlnqs" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.500458 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6475bdcbc4-8m42n" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.569919 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6b855b4fc4-jrwl8" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.805258 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-m9xms" Jan 29 12:23:53 crc kubenswrapper[4993]: I0129 12:23:53.998330 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-56b5dc77fd-v57np" Jan 29 12:24:03 crc kubenswrapper[4993]: I0129 12:24:03.778857 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" event={"ID":"66e97264-1497-4d2d-968f-e2edc9dc9017","Type":"ContainerStarted","Data":"8bb92af46fdf74c980352738e29d62bcd6d4d1389a9b1f440be69adf69f4ab46"} Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.267056 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.276180 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2b0bb-dbda-49d5-92f6-567703290c05-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs\" (UID: \"30c2b0bb-dbda-49d5-92f6-567703290c05\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.449488 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.573573 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.573673 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.579937 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-metrics-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.580840 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6c7a911f-47c8-4fab-adb0-7d179c090b29-webhook-certs\") pod \"openstack-operator-controller-manager-65dc8f5954-7dnzd\" (UID: \"6c7a911f-47c8-4fab-adb0-7d179c090b29\") " pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:05 crc kubenswrapper[4993]: I0129 12:24:05.764961 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.185554 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.185996 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkvjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b964cf4cd-2xcgn_openstack-operators(66cb8608-8dd0-4304-9245-8c4468d7c6eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.187242 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podUID="66cb8608-8dd0-4304-9245-8c4468d7c6eb" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.458530 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.458762 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p95c4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6f7455757b-ls4r8_openstack-operators(0a0fde08-a591-4fb8-a872-c0a4cd7e2758): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.459928 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podUID="0a0fde08-a591-4fb8-a872-c0a4cd7e2758" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.944156 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.944991 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9cg42,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5ccd5b7f8f-c6gdt_openstack-operators(aa36f54c-d314-4086-a64e-acec3c80c8bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 29 12:24:08 crc kubenswrapper[4993]: E0129 12:24:08.946213 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podUID="aa36f54c-d314-4086-a64e-acec3c80c8bd" Jan 29 12:24:09 crc kubenswrapper[4993]: E0129 12:24:09.662154 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b" Jan 29 12:24:09 crc kubenswrapper[4993]: E0129 12:24:09.662417 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-997f9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-c95fd9dc5-fqltq_openstack-operators(c3df3d31-0166-4aaa-b6a8-98865bc4d99c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 29 12:24:09 crc kubenswrapper[4993]: E0129 12:24:09.663606 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podUID="c3df3d31-0166-4aaa-b6a8-98865bc4d99c" Jan 29 12:24:11 crc kubenswrapper[4993]: E0129 12:24:11.310413 4993 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 29 12:24:11 crc kubenswrapper[4993]: E0129 12:24:11.310953 4993 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lqr7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-pxqq5_openstack-operators(181f3d8c-9fd7-4df6-806a-8d04ac0ff597): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 29 12:24:11 crc kubenswrapper[4993]: E0129 12:24:11.312229 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podUID="181f3d8c-9fd7-4df6-806a-8d04ac0ff597" Jan 29 12:24:12 crc kubenswrapper[4993]: I0129 12:24:12.328535 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd"] Jan 29 12:24:12 crc kubenswrapper[4993]: W0129 12:24:12.614716 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c7a911f_47c8_4fab_adb0_7d179c090b29.slice/crio-6c4bbde02cf250b4466096876e3a99836f98730e768e342431f3c77c4987a46e WatchSource:0}: Error finding container 6c4bbde02cf250b4466096876e3a99836f98730e768e342431f3c77c4987a46e: Status 404 returned error can't find the container with id 6c4bbde02cf250b4466096876e3a99836f98730e768e342431f3c77c4987a46e Jan 29 12:24:12 crc kubenswrapper[4993]: I0129 12:24:12.841886 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" event={"ID":"6c7a911f-47c8-4fab-adb0-7d179c090b29","Type":"ContainerStarted","Data":"6c4bbde02cf250b4466096876e3a99836f98730e768e342431f3c77c4987a46e"} Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.026758 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs"] Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.860431 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" event={"ID":"10455fa4-3adf-4e84-8a37-ae1241d62df6","Type":"ContainerStarted","Data":"62f9b0f28c15e4db4f28261f4c3a939f1c4fb2f2c8757d413cec48691cfe9e45"} Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.860967 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.862707 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" event={"ID":"6c7a911f-47c8-4fab-adb0-7d179c090b29","Type":"ContainerStarted","Data":"5cde18f1f83b25a4214c2c2040ff1ba69613cb7da5d9a1a8527bdfd5fd57b1b3"} Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.862881 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.866590 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" event={"ID":"30c2b0bb-dbda-49d5-92f6-567703290c05","Type":"ContainerStarted","Data":"4082d512d5a9980ce4f9e95dcd49d168bcfbfeadca14822d8a1fc0a9700c9381"} Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.868974 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" event={"ID":"66e97264-1497-4d2d-968f-e2edc9dc9017","Type":"ContainerStarted","Data":"ecdef8be327721f99453528ab15ff385fbcd297e45cce9c05a87ba234413a2b6"} Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.869128 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.883627 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" podStartSLOduration=4.190612747 podStartE2EDuration="41.88361082s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.953703938 +0000 UTC m=+1098.968834064" lastFinishedPulling="2026-01-29 12:24:12.646702011 +0000 UTC m=+1136.661832137" observedRunningTime="2026-01-29 12:24:13.877058692 +0000 UTC m=+1137.892188828" watchObservedRunningTime="2026-01-29 12:24:13.88361082 +0000 UTC m=+1137.898740946" Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.904527 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" podStartSLOduration=32.223635118 podStartE2EDuration="41.904478367s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:24:03.300441331 +0000 UTC m=+1127.315571457" lastFinishedPulling="2026-01-29 12:24:12.98128458 +0000 UTC m=+1136.996414706" observedRunningTime="2026-01-29 12:24:13.896446509 +0000 UTC m=+1137.911576635" watchObservedRunningTime="2026-01-29 12:24:13.904478367 +0000 UTC m=+1137.919608493" Jan 29 12:24:13 crc kubenswrapper[4993]: I0129 12:24:13.937990 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" podStartSLOduration=40.937969397 podStartE2EDuration="40.937969397s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:24:13.932099057 +0000 UTC m=+1137.947229223" watchObservedRunningTime="2026-01-29 12:24:13.937969397 +0000 UTC m=+1137.953099523" Jan 29 12:24:15 crc kubenswrapper[4993]: I0129 12:24:15.899789 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" event={"ID":"30c2b0bb-dbda-49d5-92f6-567703290c05","Type":"ContainerStarted","Data":"7535c98608f214de97ee2e165b36045b73554ae5ce56d5e64b60dec54a530b91"} Jan 29 12:24:15 crc kubenswrapper[4993]: I0129 12:24:15.900403 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:24:15 crc kubenswrapper[4993]: I0129 12:24:15.938337 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" podStartSLOduration=41.679239238 podStartE2EDuration="43.938316933s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:24:13.035891023 +0000 UTC m=+1137.051021149" lastFinishedPulling="2026-01-29 12:24:15.294968718 +0000 UTC m=+1139.310098844" observedRunningTime="2026-01-29 12:24:15.935089696 +0000 UTC m=+1139.950219832" watchObservedRunningTime="2026-01-29 12:24:15.938316933 +0000 UTC m=+1139.953447069" Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.156698 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.157014 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.157062 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.157721 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.157779 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47" gracePeriod=600 Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.923488 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47" exitCode=0 Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.924093 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47"} Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.924125 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217"} Jan 29 12:24:18 crc kubenswrapper[4993]: I0129 12:24:18.924146 4993 scope.go:117] "RemoveContainer" containerID="9bd05c5cd56b13d50ddad7924a1491c2dd1a4cd96424b60956f47fc1781db141" Jan 29 12:24:19 crc kubenswrapper[4993]: I0129 12:24:19.153420 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c9n87" Jan 29 12:24:19 crc kubenswrapper[4993]: E0129 12:24:19.193812 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/swift-operator@sha256:4dfb3cd42806f7989d962e2346a58c6358e70cf95c41b4890e26cb5219805ac8\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podUID="0a0fde08-a591-4fb8-a872-c0a4cd7e2758" Jan 29 12:24:21 crc kubenswrapper[4993]: E0129 12:24:21.192349 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/nova-operator@sha256:5a7235a96194f43fbbbee4085b28e1749733862ce801ef67413f496a1e5826bc\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podUID="aa36f54c-d314-4086-a64e-acec3c80c8bd" Jan 29 12:24:22 crc kubenswrapper[4993]: E0129 12:24:22.191426 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podUID="66cb8608-8dd0-4304-9245-8c4468d7c6eb" Jan 29 12:24:22 crc kubenswrapper[4993]: E0129 12:24:22.191528 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podUID="181f3d8c-9fd7-4df6-806a-8d04ac0ff597" Jan 29 12:24:23 crc kubenswrapper[4993]: E0129 12:24:23.192178 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:3e0713a6e9097420ebc622a70d44fcc5e5e9b9e036babe2966afd7c6fb5dc40b\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podUID="c3df3d31-0166-4aaa-b6a8-98865bc4d99c" Jan 29 12:24:23 crc kubenswrapper[4993]: I0129 12:24:23.588037 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-sqf57" Jan 29 12:24:25 crc kubenswrapper[4993]: I0129 12:24:25.456338 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs" Jan 29 12:24:25 crc kubenswrapper[4993]: I0129 12:24:25.770586 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-65dc8f5954-7dnzd" Jan 29 12:24:31 crc kubenswrapper[4993]: I0129 12:24:31.193124 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:24:33 crc kubenswrapper[4993]: I0129 12:24:33.018496 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" event={"ID":"0a0fde08-a591-4fb8-a872-c0a4cd7e2758","Type":"ContainerStarted","Data":"ad3c3bef6a89b625ea65913239992b58d4a7d76e74f0171d6d0560c753e8fbff"} Jan 29 12:24:33 crc kubenswrapper[4993]: I0129 12:24:33.018956 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:24:34 crc kubenswrapper[4993]: I0129 12:24:34.026660 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" event={"ID":"aa36f54c-d314-4086-a64e-acec3c80c8bd","Type":"ContainerStarted","Data":"414ee6cc2a1a4f8e9fc02e7bd9d0cc4de895109a93a7d70df05b286fbc67b58c"} Jan 29 12:24:34 crc kubenswrapper[4993]: I0129 12:24:34.026970 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:24:34 crc kubenswrapper[4993]: I0129 12:24:34.047402 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" podStartSLOduration=3.881666865 podStartE2EDuration="1m2.047384094s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.934483256 +0000 UTC m=+1098.949613372" lastFinishedPulling="2026-01-29 12:24:33.100200475 +0000 UTC m=+1157.115330601" observedRunningTime="2026-01-29 12:24:34.04278454 +0000 UTC m=+1158.057914666" watchObservedRunningTime="2026-01-29 12:24:34.047384094 +0000 UTC m=+1158.062514220" Jan 29 12:24:34 crc kubenswrapper[4993]: I0129 12:24:34.048688 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" podStartSLOduration=3.560551636 podStartE2EDuration="1m1.04868092s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.953486631 +0000 UTC m=+1098.968616767" lastFinishedPulling="2026-01-29 12:24:32.441615925 +0000 UTC m=+1156.456746051" observedRunningTime="2026-01-29 12:24:33.03855186 +0000 UTC m=+1157.053681986" watchObservedRunningTime="2026-01-29 12:24:34.04868092 +0000 UTC m=+1158.063811046" Jan 29 12:24:36 crc kubenswrapper[4993]: I0129 12:24:36.040891 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" event={"ID":"181f3d8c-9fd7-4df6-806a-8d04ac0ff597","Type":"ContainerStarted","Data":"f5079926c2066cf1e258d8e676b0002b5d1a253feb1f1368f661bfef31cd4302"} Jan 29 12:24:36 crc kubenswrapper[4993]: I0129 12:24:36.066050 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-pxqq5" podStartSLOduration=3.086558491 podStartE2EDuration="1m3.066030709s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.927781974 +0000 UTC m=+1098.942912100" lastFinishedPulling="2026-01-29 12:24:34.907254192 +0000 UTC m=+1158.922384318" observedRunningTime="2026-01-29 12:24:36.057167127 +0000 UTC m=+1160.072297263" watchObservedRunningTime="2026-01-29 12:24:36.066030709 +0000 UTC m=+1160.081160835" Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.061321 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" event={"ID":"c3df3d31-0166-4aaa-b6a8-98865bc4d99c","Type":"ContainerStarted","Data":"55dcdd95165e6f91ec452579e914af2ef639a6a9021f50d8fa8adc8dd5a0c156"} Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.062073 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.062787 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" event={"ID":"66cb8608-8dd0-4304-9245-8c4468d7c6eb","Type":"ContainerStarted","Data":"277e28e1ae9311625d3c0f0f857b34bbbdde6b48b3306d9da08524e8c36bc8cd"} Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.062987 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.078682 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" podStartSLOduration=2.287961499 podStartE2EDuration="1m6.078652592s" podCreationTimestamp="2026-01-29 12:23:33 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.948374025 +0000 UTC m=+1098.963504151" lastFinishedPulling="2026-01-29 12:24:38.739065118 +0000 UTC m=+1162.754195244" observedRunningTime="2026-01-29 12:24:39.076082583 +0000 UTC m=+1163.091212709" watchObservedRunningTime="2026-01-29 12:24:39.078652592 +0000 UTC m=+1163.093782718" Jan 29 12:24:39 crc kubenswrapper[4993]: I0129 12:24:39.096974 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" podStartSLOduration=3.97480082 podStartE2EDuration="1m7.096951589s" podCreationTimestamp="2026-01-29 12:23:32 +0000 UTC" firstStartedPulling="2026-01-29 12:23:34.945116751 +0000 UTC m=+1098.960246887" lastFinishedPulling="2026-01-29 12:24:38.06726753 +0000 UTC m=+1162.082397656" observedRunningTime="2026-01-29 12:24:39.091056029 +0000 UTC m=+1163.106186185" watchObservedRunningTime="2026-01-29 12:24:39.096951589 +0000 UTC m=+1163.112081715" Jan 29 12:24:43 crc kubenswrapper[4993]: I0129 12:24:43.553346 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5ccd5b7f8f-c6gdt" Jan 29 12:24:43 crc kubenswrapper[4993]: I0129 12:24:43.641096 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-2xcgn" Jan 29 12:24:43 crc kubenswrapper[4993]: I0129 12:24:43.680342 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-ls4r8" Jan 29 12:24:43 crc kubenswrapper[4993]: I0129 12:24:43.709478 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-c95fd9dc5-fqltq" Jan 29 12:26:18 crc kubenswrapper[4993]: I0129 12:26:18.156073 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:26:18 crc kubenswrapper[4993]: I0129 12:26:18.156679 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:26:48 crc kubenswrapper[4993]: I0129 12:26:48.155897 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:26:48 crc kubenswrapper[4993]: I0129 12:26:48.156511 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:27:18 crc kubenswrapper[4993]: I0129 12:27:18.156673 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:27:18 crc kubenswrapper[4993]: I0129 12:27:18.157247 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:27:18 crc kubenswrapper[4993]: I0129 12:27:18.157301 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:27:18 crc kubenswrapper[4993]: I0129 12:27:18.157898 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:27:18 crc kubenswrapper[4993]: I0129 12:27:18.157954 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217" gracePeriod=600 Jan 29 12:27:19 crc kubenswrapper[4993]: I0129 12:27:19.263636 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217" exitCode=0 Jan 29 12:27:19 crc kubenswrapper[4993]: I0129 12:27:19.263718 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217"} Jan 29 12:27:19 crc kubenswrapper[4993]: I0129 12:27:19.263940 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1"} Jan 29 12:27:19 crc kubenswrapper[4993]: I0129 12:27:19.263961 4993 scope.go:117] "RemoveContainer" containerID="0d477cf92d5ef97ad15b62408bdf2b06aa98e5c9cb065ec54d804304fd202b47" Jan 29 12:29:18 crc kubenswrapper[4993]: I0129 12:29:18.155872 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:29:18 crc kubenswrapper[4993]: I0129 12:29:18.156415 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:29:48 crc kubenswrapper[4993]: I0129 12:29:48.155973 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:29:48 crc kubenswrapper[4993]: I0129 12:29:48.156633 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.038707 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.040567 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.049253 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.134442 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpfk\" (UniqueName: \"kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.134781 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.134934 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.236351 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpfk\" (UniqueName: \"kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.236467 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.236534 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.237873 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.238266 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.272395 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpfk\" (UniqueName: \"kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk\") pod \"redhat-operators-gjmqq\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.361488 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:29:55 crc kubenswrapper[4993]: I0129 12:29:55.615226 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:29:56 crc kubenswrapper[4993]: I0129 12:29:56.431514 4993 generic.go:334] "Generic (PLEG): container finished" podID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerID="b391c29811f94865ae771b1f8e95e7fa3f296f0788faf0295b2094dc10c31487" exitCode=0 Jan 29 12:29:56 crc kubenswrapper[4993]: I0129 12:29:56.431572 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerDied","Data":"b391c29811f94865ae771b1f8e95e7fa3f296f0788faf0295b2094dc10c31487"} Jan 29 12:29:56 crc kubenswrapper[4993]: I0129 12:29:56.431605 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerStarted","Data":"8ee797c75cd75f80b69e94c4ceeadab3f7b6860e671ea2668a7231f1701f5962"} Jan 29 12:29:56 crc kubenswrapper[4993]: I0129 12:29:56.433563 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:29:58 crc kubenswrapper[4993]: I0129 12:29:58.446680 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerStarted","Data":"5bae09d04684d2c91c9de1bea0fe4aca4f5e3e050ff36bf7ab3b76bc3ecb03f6"} Jan 29 12:29:59 crc kubenswrapper[4993]: I0129 12:29:59.455332 4993 generic.go:334] "Generic (PLEG): container finished" podID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerID="5bae09d04684d2c91c9de1bea0fe4aca4f5e3e050ff36bf7ab3b76bc3ecb03f6" exitCode=0 Jan 29 12:29:59 crc kubenswrapper[4993]: I0129 12:29:59.455394 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerDied","Data":"5bae09d04684d2c91c9de1bea0fe4aca4f5e3e050ff36bf7ab3b76bc3ecb03f6"} Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.146873 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb"] Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.148494 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.151542 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.154432 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb"] Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.156472 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.210564 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.210655 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.210713 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh9qw\" (UniqueName: \"kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.311987 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh9qw\" (UniqueName: \"kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.312103 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.312157 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.313584 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.318588 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.329842 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh9qw\" (UniqueName: \"kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw\") pod \"collect-profiles-29494830-bbbxb\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.489659 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerStarted","Data":"0155a8a64f7cbff3b37bd63499aa24bc105bc080250227b6e677d4619f8f8ecb"} Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.511359 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.515303 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gjmqq" podStartSLOduration=1.838385717 podStartE2EDuration="5.515284472s" podCreationTimestamp="2026-01-29 12:29:55 +0000 UTC" firstStartedPulling="2026-01-29 12:29:56.433332044 +0000 UTC m=+1480.448462170" lastFinishedPulling="2026-01-29 12:30:00.110230789 +0000 UTC m=+1484.125360925" observedRunningTime="2026-01-29 12:30:00.506963825 +0000 UTC m=+1484.522093961" watchObservedRunningTime="2026-01-29 12:30:00.515284472 +0000 UTC m=+1484.530414598" Jan 29 12:30:00 crc kubenswrapper[4993]: I0129 12:30:00.931094 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb"] Jan 29 12:30:01 crc kubenswrapper[4993]: I0129 12:30:01.500155 4993 generic.go:334] "Generic (PLEG): container finished" podID="beb4a452-d27e-4637-af3f-682ae46b0871" containerID="217cf8be7340e063ff3ba7c0593f6333b189b1eb9a5fcacd7770d7287da48526" exitCode=0 Jan 29 12:30:01 crc kubenswrapper[4993]: I0129 12:30:01.500234 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" event={"ID":"beb4a452-d27e-4637-af3f-682ae46b0871","Type":"ContainerDied","Data":"217cf8be7340e063ff3ba7c0593f6333b189b1eb9a5fcacd7770d7287da48526"} Jan 29 12:30:01 crc kubenswrapper[4993]: I0129 12:30:01.500663 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" event={"ID":"beb4a452-d27e-4637-af3f-682ae46b0871","Type":"ContainerStarted","Data":"524ee58cfae354b40729708f34fcf372136b3447ee21df463571f615d141b29c"} Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.834093 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.957011 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume\") pod \"beb4a452-d27e-4637-af3f-682ae46b0871\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.957069 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh9qw\" (UniqueName: \"kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw\") pod \"beb4a452-d27e-4637-af3f-682ae46b0871\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.957125 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume\") pod \"beb4a452-d27e-4637-af3f-682ae46b0871\" (UID: \"beb4a452-d27e-4637-af3f-682ae46b0871\") " Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.957876 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume" (OuterVolumeSpecName: "config-volume") pod "beb4a452-d27e-4637-af3f-682ae46b0871" (UID: "beb4a452-d27e-4637-af3f-682ae46b0871"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.963762 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "beb4a452-d27e-4637-af3f-682ae46b0871" (UID: "beb4a452-d27e-4637-af3f-682ae46b0871"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:30:02 crc kubenswrapper[4993]: I0129 12:30:02.964390 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw" (OuterVolumeSpecName: "kube-api-access-qh9qw") pod "beb4a452-d27e-4637-af3f-682ae46b0871" (UID: "beb4a452-d27e-4637-af3f-682ae46b0871"). InnerVolumeSpecName "kube-api-access-qh9qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.059113 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/beb4a452-d27e-4637-af3f-682ae46b0871-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.059166 4993 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/beb4a452-d27e-4637-af3f-682ae46b0871-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.059201 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh9qw\" (UniqueName: \"kubernetes.io/projected/beb4a452-d27e-4637-af3f-682ae46b0871-kube-api-access-qh9qw\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.514316 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" event={"ID":"beb4a452-d27e-4637-af3f-682ae46b0871","Type":"ContainerDied","Data":"524ee58cfae354b40729708f34fcf372136b3447ee21df463571f615d141b29c"} Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.514363 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="524ee58cfae354b40729708f34fcf372136b3447ee21df463571f615d141b29c" Jan 29 12:30:03 crc kubenswrapper[4993]: I0129 12:30:03.514381 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494830-bbbxb" Jan 29 12:30:05 crc kubenswrapper[4993]: I0129 12:30:05.362145 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:05 crc kubenswrapper[4993]: I0129 12:30:05.362408 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:06 crc kubenswrapper[4993]: I0129 12:30:06.408136 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gjmqq" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="registry-server" probeResult="failure" output=< Jan 29 12:30:06 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:30:06 crc kubenswrapper[4993]: > Jan 29 12:30:15 crc kubenswrapper[4993]: I0129 12:30:15.405011 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:15 crc kubenswrapper[4993]: I0129 12:30:15.462882 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:15 crc kubenswrapper[4993]: I0129 12:30:15.645061 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:30:16 crc kubenswrapper[4993]: I0129 12:30:16.612890 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gjmqq" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="registry-server" containerID="cri-o://0155a8a64f7cbff3b37bd63499aa24bc105bc080250227b6e677d4619f8f8ecb" gracePeriod=2 Jan 29 12:30:18 crc kubenswrapper[4993]: I0129 12:30:18.156458 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:30:18 crc kubenswrapper[4993]: I0129 12:30:18.156957 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:30:18 crc kubenswrapper[4993]: I0129 12:30:18.157415 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:30:18 crc kubenswrapper[4993]: I0129 12:30:18.158274 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:30:18 crc kubenswrapper[4993]: I0129 12:30:18.158367 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" gracePeriod=600 Jan 29 12:30:19 crc kubenswrapper[4993]: E0129 12:30:19.449905 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.636602 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" exitCode=0 Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.636683 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1"} Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.636735 4993 scope.go:117] "RemoveContainer" containerID="13608b28325d9560de7e35f655ca63405716b2e17a31208100f9ac5689846217" Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.637376 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:30:19 crc kubenswrapper[4993]: E0129 12:30:19.637633 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.641792 4993 generic.go:334] "Generic (PLEG): container finished" podID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerID="0155a8a64f7cbff3b37bd63499aa24bc105bc080250227b6e677d4619f8f8ecb" exitCode=0 Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.641841 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerDied","Data":"0155a8a64f7cbff3b37bd63499aa24bc105bc080250227b6e677d4619f8f8ecb"} Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.786756 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.901769 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content\") pod \"bdb4c08a-948f-428f-bf8e-1689210e7df4\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.901929 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities\") pod \"bdb4c08a-948f-428f-bf8e-1689210e7df4\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.902025 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddpfk\" (UniqueName: \"kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk\") pod \"bdb4c08a-948f-428f-bf8e-1689210e7df4\" (UID: \"bdb4c08a-948f-428f-bf8e-1689210e7df4\") " Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.903099 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities" (OuterVolumeSpecName: "utilities") pod "bdb4c08a-948f-428f-bf8e-1689210e7df4" (UID: "bdb4c08a-948f-428f-bf8e-1689210e7df4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:30:19 crc kubenswrapper[4993]: I0129 12:30:19.908461 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk" (OuterVolumeSpecName: "kube-api-access-ddpfk") pod "bdb4c08a-948f-428f-bf8e-1689210e7df4" (UID: "bdb4c08a-948f-428f-bf8e-1689210e7df4"). InnerVolumeSpecName "kube-api-access-ddpfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.004023 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddpfk\" (UniqueName: \"kubernetes.io/projected/bdb4c08a-948f-428f-bf8e-1689210e7df4-kube-api-access-ddpfk\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.004050 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.039713 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdb4c08a-948f-428f-bf8e-1689210e7df4" (UID: "bdb4c08a-948f-428f-bf8e-1689210e7df4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.105831 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdb4c08a-948f-428f-bf8e-1689210e7df4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.653963 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gjmqq" event={"ID":"bdb4c08a-948f-428f-bf8e-1689210e7df4","Type":"ContainerDied","Data":"8ee797c75cd75f80b69e94c4ceeadab3f7b6860e671ea2668a7231f1701f5962"} Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.654736 4993 scope.go:117] "RemoveContainer" containerID="0155a8a64f7cbff3b37bd63499aa24bc105bc080250227b6e677d4619f8f8ecb" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.654095 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gjmqq" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.680407 4993 scope.go:117] "RemoveContainer" containerID="5bae09d04684d2c91c9de1bea0fe4aca4f5e3e050ff36bf7ab3b76bc3ecb03f6" Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.693040 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.699812 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gjmqq"] Jan 29 12:30:20 crc kubenswrapper[4993]: I0129 12:30:20.710628 4993 scope.go:117] "RemoveContainer" containerID="b391c29811f94865ae771b1f8e95e7fa3f296f0788faf0295b2094dc10c31487" Jan 29 12:30:21 crc kubenswrapper[4993]: I0129 12:30:21.199682 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" path="/var/lib/kubelet/pods/bdb4c08a-948f-428f-bf8e-1689210e7df4/volumes" Jan 29 12:30:30 crc kubenswrapper[4993]: I0129 12:30:30.190998 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:30:30 crc kubenswrapper[4993]: E0129 12:30:30.191544 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:30:42 crc kubenswrapper[4993]: I0129 12:30:42.191895 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:30:42 crc kubenswrapper[4993]: E0129 12:30:42.193217 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:30:57 crc kubenswrapper[4993]: I0129 12:30:57.197582 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:30:57 crc kubenswrapper[4993]: E0129 12:30:57.198442 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.955446 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:02 crc kubenswrapper[4993]: E0129 12:31:02.956486 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="registry-server" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956518 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="registry-server" Jan 29 12:31:02 crc kubenswrapper[4993]: E0129 12:31:02.956548 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="extract-content" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956556 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="extract-content" Jan 29 12:31:02 crc kubenswrapper[4993]: E0129 12:31:02.956568 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="extract-utilities" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956577 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="extract-utilities" Jan 29 12:31:02 crc kubenswrapper[4993]: E0129 12:31:02.956590 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb4a452-d27e-4637-af3f-682ae46b0871" containerName="collect-profiles" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956598 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb4a452-d27e-4637-af3f-682ae46b0871" containerName="collect-profiles" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956758 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb4a452-d27e-4637-af3f-682ae46b0871" containerName="collect-profiles" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.956787 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb4c08a-948f-428f-bf8e-1689210e7df4" containerName="registry-server" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.958039 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:02 crc kubenswrapper[4993]: I0129 12:31:02.966741 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.083451 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.083568 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5zdd\" (UniqueName: \"kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.083609 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.184784 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.184850 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5zdd\" (UniqueName: \"kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.184901 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.185437 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.185708 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.209646 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5zdd\" (UniqueName: \"kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd\") pod \"redhat-marketplace-qgnzr\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.282214 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.754354 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:03 crc kubenswrapper[4993]: I0129 12:31:03.975381 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerStarted","Data":"cd0a047599162951d9852b2552c5adf141dd97fb1fb49700379dbb48ae35676e"} Jan 29 12:31:04 crc kubenswrapper[4993]: I0129 12:31:04.984312 4993 generic.go:334] "Generic (PLEG): container finished" podID="c103e05a-98e3-4329-a417-e2132e086f9e" containerID="5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d" exitCode=0 Jan 29 12:31:04 crc kubenswrapper[4993]: I0129 12:31:04.984382 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerDied","Data":"5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d"} Jan 29 12:31:05 crc kubenswrapper[4993]: I0129 12:31:05.993839 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerStarted","Data":"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45"} Jan 29 12:31:07 crc kubenswrapper[4993]: I0129 12:31:07.001072 4993 generic.go:334] "Generic (PLEG): container finished" podID="c103e05a-98e3-4329-a417-e2132e086f9e" containerID="3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45" exitCode=0 Jan 29 12:31:07 crc kubenswrapper[4993]: I0129 12:31:07.001153 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerDied","Data":"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45"} Jan 29 12:31:08 crc kubenswrapper[4993]: I0129 12:31:08.010546 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerStarted","Data":"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e"} Jan 29 12:31:08 crc kubenswrapper[4993]: I0129 12:31:08.033651 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qgnzr" podStartSLOduration=3.618351974 podStartE2EDuration="6.033636925s" podCreationTimestamp="2026-01-29 12:31:02 +0000 UTC" firstStartedPulling="2026-01-29 12:31:04.986123295 +0000 UTC m=+1549.001253421" lastFinishedPulling="2026-01-29 12:31:07.401408246 +0000 UTC m=+1551.416538372" observedRunningTime="2026-01-29 12:31:08.032593637 +0000 UTC m=+1552.047723763" watchObservedRunningTime="2026-01-29 12:31:08.033636925 +0000 UTC m=+1552.048767051" Jan 29 12:31:09 crc kubenswrapper[4993]: I0129 12:31:09.189991 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:31:09 crc kubenswrapper[4993]: E0129 12:31:09.190479 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:31:13 crc kubenswrapper[4993]: I0129 12:31:13.283175 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:13 crc kubenswrapper[4993]: I0129 12:31:13.283525 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:13 crc kubenswrapper[4993]: I0129 12:31:13.350039 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:14 crc kubenswrapper[4993]: I0129 12:31:14.098015 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:14 crc kubenswrapper[4993]: I0129 12:31:14.145372 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.067916 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qgnzr" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="registry-server" containerID="cri-o://a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e" gracePeriod=2 Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.478466 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.581259 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5zdd\" (UniqueName: \"kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd\") pod \"c103e05a-98e3-4329-a417-e2132e086f9e\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.581323 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities\") pod \"c103e05a-98e3-4329-a417-e2132e086f9e\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.581378 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content\") pod \"c103e05a-98e3-4329-a417-e2132e086f9e\" (UID: \"c103e05a-98e3-4329-a417-e2132e086f9e\") " Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.582600 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities" (OuterVolumeSpecName: "utilities") pod "c103e05a-98e3-4329-a417-e2132e086f9e" (UID: "c103e05a-98e3-4329-a417-e2132e086f9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.586973 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd" (OuterVolumeSpecName: "kube-api-access-x5zdd") pod "c103e05a-98e3-4329-a417-e2132e086f9e" (UID: "c103e05a-98e3-4329-a417-e2132e086f9e"). InnerVolumeSpecName "kube-api-access-x5zdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.683725 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5zdd\" (UniqueName: \"kubernetes.io/projected/c103e05a-98e3-4329-a417-e2132e086f9e-kube-api-access-x5zdd\") on node \"crc\" DevicePath \"\"" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.683754 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.732043 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c103e05a-98e3-4329-a417-e2132e086f9e" (UID: "c103e05a-98e3-4329-a417-e2132e086f9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:31:16 crc kubenswrapper[4993]: I0129 12:31:16.786232 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c103e05a-98e3-4329-a417-e2132e086f9e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.076445 4993 generic.go:334] "Generic (PLEG): container finished" podID="c103e05a-98e3-4329-a417-e2132e086f9e" containerID="a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e" exitCode=0 Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.076492 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerDied","Data":"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e"} Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.076799 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgnzr" event={"ID":"c103e05a-98e3-4329-a417-e2132e086f9e","Type":"ContainerDied","Data":"cd0a047599162951d9852b2552c5adf141dd97fb1fb49700379dbb48ae35676e"} Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.076826 4993 scope.go:117] "RemoveContainer" containerID="a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.076513 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgnzr" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.100178 4993 scope.go:117] "RemoveContainer" containerID="3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.117846 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.126579 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgnzr"] Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.138276 4993 scope.go:117] "RemoveContainer" containerID="5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.158849 4993 scope.go:117] "RemoveContainer" containerID="a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e" Jan 29 12:31:17 crc kubenswrapper[4993]: E0129 12:31:17.159419 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e\": container with ID starting with a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e not found: ID does not exist" containerID="a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.159500 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e"} err="failed to get container status \"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e\": rpc error: code = NotFound desc = could not find container \"a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e\": container with ID starting with a9bc8cec7757e3cbdcf08845e87c2f0bfd4c8ca5de2a71c572da6326a3009a7e not found: ID does not exist" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.159560 4993 scope.go:117] "RemoveContainer" containerID="3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45" Jan 29 12:31:17 crc kubenswrapper[4993]: E0129 12:31:17.160547 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45\": container with ID starting with 3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45 not found: ID does not exist" containerID="3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.160583 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45"} err="failed to get container status \"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45\": rpc error: code = NotFound desc = could not find container \"3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45\": container with ID starting with 3e7509d174661331be3c89295876982029cdd3392cfd0bf7b1e9d8277ab28e45 not found: ID does not exist" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.160609 4993 scope.go:117] "RemoveContainer" containerID="5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d" Jan 29 12:31:17 crc kubenswrapper[4993]: E0129 12:31:17.161503 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d\": container with ID starting with 5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d not found: ID does not exist" containerID="5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.161565 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d"} err="failed to get container status \"5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d\": rpc error: code = NotFound desc = could not find container \"5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d\": container with ID starting with 5712e9646f933ea4fc3ea803075f94218ea9af16066289041e8fd06de83e286d not found: ID does not exist" Jan 29 12:31:17 crc kubenswrapper[4993]: I0129 12:31:17.201505 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" path="/var/lib/kubelet/pods/c103e05a-98e3-4329-a417-e2132e086f9e/volumes" Jan 29 12:31:23 crc kubenswrapper[4993]: I0129 12:31:23.190875 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:31:23 crc kubenswrapper[4993]: E0129 12:31:23.191662 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:31:36 crc kubenswrapper[4993]: I0129 12:31:36.190238 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:31:36 crc kubenswrapper[4993]: E0129 12:31:36.191032 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:31:48 crc kubenswrapper[4993]: I0129 12:31:48.190538 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:31:48 crc kubenswrapper[4993]: E0129 12:31:48.191138 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:31:59 crc kubenswrapper[4993]: I0129 12:31:59.190847 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:31:59 crc kubenswrapper[4993]: E0129 12:31:59.191765 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:32:11 crc kubenswrapper[4993]: I0129 12:32:11.190845 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:32:11 crc kubenswrapper[4993]: E0129 12:32:11.191584 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.670408 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:21 crc kubenswrapper[4993]: E0129 12:32:21.671204 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="extract-content" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.671216 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="extract-content" Jan 29 12:32:21 crc kubenswrapper[4993]: E0129 12:32:21.671225 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="registry-server" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.671231 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="registry-server" Jan 29 12:32:21 crc kubenswrapper[4993]: E0129 12:32:21.671249 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="extract-utilities" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.671256 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="extract-utilities" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.671388 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="c103e05a-98e3-4329-a417-e2132e086f9e" containerName="registry-server" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.672400 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.685233 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.841634 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.841727 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpq4\" (UniqueName: \"kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.841773 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.943133 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpq4\" (UniqueName: \"kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.943568 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.943767 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.944106 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.944222 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:21 crc kubenswrapper[4993]: I0129 12:32:21.965277 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpq4\" (UniqueName: \"kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4\") pod \"community-operators-lzw7w\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:22 crc kubenswrapper[4993]: I0129 12:32:22.043066 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:22 crc kubenswrapper[4993]: I0129 12:32:22.287964 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:22 crc kubenswrapper[4993]: I0129 12:32:22.564133 4993 generic.go:334] "Generic (PLEG): container finished" podID="662ca330-c59e-49c4-befd-39ff89161599" containerID="66fbec9c3133225da6a5377b932ceb62379d5fc3d69fc83f1c490c591f2c8d4d" exitCode=0 Jan 29 12:32:22 crc kubenswrapper[4993]: I0129 12:32:22.564392 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerDied","Data":"66fbec9c3133225da6a5377b932ceb62379d5fc3d69fc83f1c490c591f2c8d4d"} Jan 29 12:32:22 crc kubenswrapper[4993]: I0129 12:32:22.564525 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerStarted","Data":"2c3e7edc30d6e58c513d8f999878d24196fa613c02b62d489c8fd6f3fe7701ae"} Jan 29 12:32:23 crc kubenswrapper[4993]: I0129 12:32:23.575043 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerStarted","Data":"e32be417ff3fed0a4a8900eba69dceadccce8887f944b46ceef4734db09bb83c"} Jan 29 12:32:24 crc kubenswrapper[4993]: I0129 12:32:24.189948 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:32:24 crc kubenswrapper[4993]: E0129 12:32:24.190472 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:32:24 crc kubenswrapper[4993]: I0129 12:32:24.582506 4993 generic.go:334] "Generic (PLEG): container finished" podID="662ca330-c59e-49c4-befd-39ff89161599" containerID="e32be417ff3fed0a4a8900eba69dceadccce8887f944b46ceef4734db09bb83c" exitCode=0 Jan 29 12:32:24 crc kubenswrapper[4993]: I0129 12:32:24.582559 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerDied","Data":"e32be417ff3fed0a4a8900eba69dceadccce8887f944b46ceef4734db09bb83c"} Jan 29 12:32:25 crc kubenswrapper[4993]: I0129 12:32:25.591833 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerStarted","Data":"6bcca6985e9a9dbd28868f5c8ba9adff1117d7fb7a3ffb6eea0182504a8c9fe5"} Jan 29 12:32:25 crc kubenswrapper[4993]: I0129 12:32:25.618151 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lzw7w" podStartSLOduration=2.148428174 podStartE2EDuration="4.618130401s" podCreationTimestamp="2026-01-29 12:32:21 +0000 UTC" firstStartedPulling="2026-01-29 12:32:22.566154561 +0000 UTC m=+1626.581284687" lastFinishedPulling="2026-01-29 12:32:25.035856778 +0000 UTC m=+1629.050986914" observedRunningTime="2026-01-29 12:32:25.615794278 +0000 UTC m=+1629.630924524" watchObservedRunningTime="2026-01-29 12:32:25.618130401 +0000 UTC m=+1629.633260527" Jan 29 12:32:32 crc kubenswrapper[4993]: I0129 12:32:32.044245 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:32 crc kubenswrapper[4993]: I0129 12:32:32.044591 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:32 crc kubenswrapper[4993]: I0129 12:32:32.111449 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:32 crc kubenswrapper[4993]: I0129 12:32:32.698717 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:32 crc kubenswrapper[4993]: I0129 12:32:32.751926 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:34 crc kubenswrapper[4993]: I0129 12:32:34.662177 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lzw7w" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="registry-server" containerID="cri-o://6bcca6985e9a9dbd28868f5c8ba9adff1117d7fb7a3ffb6eea0182504a8c9fe5" gracePeriod=2 Jan 29 12:32:36 crc kubenswrapper[4993]: I0129 12:32:36.679720 4993 generic.go:334] "Generic (PLEG): container finished" podID="662ca330-c59e-49c4-befd-39ff89161599" containerID="6bcca6985e9a9dbd28868f5c8ba9adff1117d7fb7a3ffb6eea0182504a8c9fe5" exitCode=0 Jan 29 12:32:36 crc kubenswrapper[4993]: I0129 12:32:36.679823 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerDied","Data":"6bcca6985e9a9dbd28868f5c8ba9adff1117d7fb7a3ffb6eea0182504a8c9fe5"} Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.008661 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.058959 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities\") pod \"662ca330-c59e-49c4-befd-39ff89161599\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.059018 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmpq4\" (UniqueName: \"kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4\") pod \"662ca330-c59e-49c4-befd-39ff89161599\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.059064 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content\") pod \"662ca330-c59e-49c4-befd-39ff89161599\" (UID: \"662ca330-c59e-49c4-befd-39ff89161599\") " Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.059770 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities" (OuterVolumeSpecName: "utilities") pod "662ca330-c59e-49c4-befd-39ff89161599" (UID: "662ca330-c59e-49c4-befd-39ff89161599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.064256 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4" (OuterVolumeSpecName: "kube-api-access-xmpq4") pod "662ca330-c59e-49c4-befd-39ff89161599" (UID: "662ca330-c59e-49c4-befd-39ff89161599"). InnerVolumeSpecName "kube-api-access-xmpq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.107985 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "662ca330-c59e-49c4-befd-39ff89161599" (UID: "662ca330-c59e-49c4-befd-39ff89161599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.159611 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.159661 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmpq4\" (UniqueName: \"kubernetes.io/projected/662ca330-c59e-49c4-befd-39ff89161599-kube-api-access-xmpq4\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.159675 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/662ca330-c59e-49c4-befd-39ff89161599-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.194139 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:32:37 crc kubenswrapper[4993]: E0129 12:32:37.194477 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.691521 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzw7w" event={"ID":"662ca330-c59e-49c4-befd-39ff89161599","Type":"ContainerDied","Data":"2c3e7edc30d6e58c513d8f999878d24196fa613c02b62d489c8fd6f3fe7701ae"} Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.691585 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzw7w" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.691654 4993 scope.go:117] "RemoveContainer" containerID="6bcca6985e9a9dbd28868f5c8ba9adff1117d7fb7a3ffb6eea0182504a8c9fe5" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.720863 4993 scope.go:117] "RemoveContainer" containerID="e32be417ff3fed0a4a8900eba69dceadccce8887f944b46ceef4734db09bb83c" Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.722512 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.729734 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lzw7w"] Jan 29 12:32:37 crc kubenswrapper[4993]: I0129 12:32:37.743035 4993 scope.go:117] "RemoveContainer" containerID="66fbec9c3133225da6a5377b932ceb62379d5fc3d69fc83f1c490c591f2c8d4d" Jan 29 12:32:39 crc kubenswrapper[4993]: I0129 12:32:39.200120 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="662ca330-c59e-49c4-befd-39ff89161599" path="/var/lib/kubelet/pods/662ca330-c59e-49c4-befd-39ff89161599/volumes" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.539720 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:45 crc kubenswrapper[4993]: E0129 12:32:45.540643 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="registry-server" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.540663 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="registry-server" Jan 29 12:32:45 crc kubenswrapper[4993]: E0129 12:32:45.540700 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="extract-utilities" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.540709 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="extract-utilities" Jan 29 12:32:45 crc kubenswrapper[4993]: E0129 12:32:45.540720 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="extract-content" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.540729 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="extract-content" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.540886 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="662ca330-c59e-49c4-befd-39ff89161599" containerName="registry-server" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.542010 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.555685 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.592271 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b254\" (UniqueName: \"kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.592462 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.592532 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.694448 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b254\" (UniqueName: \"kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.694790 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.694923 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.695447 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.695472 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.714223 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b254\" (UniqueName: \"kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254\") pod \"certified-operators-glr7t\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:45 crc kubenswrapper[4993]: I0129 12:32:45.861504 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:46 crc kubenswrapper[4993]: I0129 12:32:46.353484 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:46 crc kubenswrapper[4993]: I0129 12:32:46.767354 4993 generic.go:334] "Generic (PLEG): container finished" podID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerID="d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6" exitCode=0 Jan 29 12:32:46 crc kubenswrapper[4993]: I0129 12:32:46.767453 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerDied","Data":"d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6"} Jan 29 12:32:46 crc kubenswrapper[4993]: I0129 12:32:46.767639 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerStarted","Data":"d4bc48fc0b25469ed049b9fdea452e05f2e6f38e6bbc039cec6fd410d0d0fd20"} Jan 29 12:32:48 crc kubenswrapper[4993]: I0129 12:32:48.783372 4993 generic.go:334] "Generic (PLEG): container finished" podID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerID="40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b" exitCode=0 Jan 29 12:32:48 crc kubenswrapper[4993]: I0129 12:32:48.783432 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerDied","Data":"40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b"} Jan 29 12:32:49 crc kubenswrapper[4993]: I0129 12:32:49.791978 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerStarted","Data":"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc"} Jan 29 12:32:49 crc kubenswrapper[4993]: I0129 12:32:49.814700 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-glr7t" podStartSLOduration=2.084515073 podStartE2EDuration="4.814674108s" podCreationTimestamp="2026-01-29 12:32:45 +0000 UTC" firstStartedPulling="2026-01-29 12:32:46.769050832 +0000 UTC m=+1650.784180948" lastFinishedPulling="2026-01-29 12:32:49.499209847 +0000 UTC m=+1653.514339983" observedRunningTime="2026-01-29 12:32:49.810264928 +0000 UTC m=+1653.825395074" watchObservedRunningTime="2026-01-29 12:32:49.814674108 +0000 UTC m=+1653.829804244" Jan 29 12:32:50 crc kubenswrapper[4993]: I0129 12:32:50.191008 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:32:50 crc kubenswrapper[4993]: E0129 12:32:50.191252 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:32:55 crc kubenswrapper[4993]: I0129 12:32:55.863380 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:55 crc kubenswrapper[4993]: I0129 12:32:55.863928 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:55 crc kubenswrapper[4993]: I0129 12:32:55.911402 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:56 crc kubenswrapper[4993]: I0129 12:32:56.932005 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:56 crc kubenswrapper[4993]: I0129 12:32:56.992148 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:58 crc kubenswrapper[4993]: I0129 12:32:58.867679 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-glr7t" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="registry-server" containerID="cri-o://88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc" gracePeriod=2 Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.260075 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.287459 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities\") pod \"ba0a684d-6f4e-4c1f-8480-6d7167750711\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.287569 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b254\" (UniqueName: \"kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254\") pod \"ba0a684d-6f4e-4c1f-8480-6d7167750711\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.287604 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content\") pod \"ba0a684d-6f4e-4c1f-8480-6d7167750711\" (UID: \"ba0a684d-6f4e-4c1f-8480-6d7167750711\") " Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.288409 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities" (OuterVolumeSpecName: "utilities") pod "ba0a684d-6f4e-4c1f-8480-6d7167750711" (UID: "ba0a684d-6f4e-4c1f-8480-6d7167750711"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.289056 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.293601 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254" (OuterVolumeSpecName: "kube-api-access-6b254") pod "ba0a684d-6f4e-4c1f-8480-6d7167750711" (UID: "ba0a684d-6f4e-4c1f-8480-6d7167750711"). InnerVolumeSpecName "kube-api-access-6b254". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.340876 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba0a684d-6f4e-4c1f-8480-6d7167750711" (UID: "ba0a684d-6f4e-4c1f-8480-6d7167750711"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.389847 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b254\" (UniqueName: \"kubernetes.io/projected/ba0a684d-6f4e-4c1f-8480-6d7167750711-kube-api-access-6b254\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.389879 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba0a684d-6f4e-4c1f-8480-6d7167750711-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.876122 4993 generic.go:334] "Generic (PLEG): container finished" podID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerID="88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc" exitCode=0 Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.876256 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glr7t" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.876257 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerDied","Data":"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc"} Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.877049 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glr7t" event={"ID":"ba0a684d-6f4e-4c1f-8480-6d7167750711","Type":"ContainerDied","Data":"d4bc48fc0b25469ed049b9fdea452e05f2e6f38e6bbc039cec6fd410d0d0fd20"} Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.877074 4993 scope.go:117] "RemoveContainer" containerID="88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.896237 4993 scope.go:117] "RemoveContainer" containerID="40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.917623 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.923339 4993 scope.go:117] "RemoveContainer" containerID="d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.924549 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-glr7t"] Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.945801 4993 scope.go:117] "RemoveContainer" containerID="88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc" Jan 29 12:32:59 crc kubenswrapper[4993]: E0129 12:32:59.946214 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc\": container with ID starting with 88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc not found: ID does not exist" containerID="88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.946266 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc"} err="failed to get container status \"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc\": rpc error: code = NotFound desc = could not find container \"88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc\": container with ID starting with 88ff0b78865d80c05bc857f166f84e824763dc27c2b7cf6887662e763befa1fc not found: ID does not exist" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.946293 4993 scope.go:117] "RemoveContainer" containerID="40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b" Jan 29 12:32:59 crc kubenswrapper[4993]: E0129 12:32:59.946606 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b\": container with ID starting with 40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b not found: ID does not exist" containerID="40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.946639 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b"} err="failed to get container status \"40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b\": rpc error: code = NotFound desc = could not find container \"40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b\": container with ID starting with 40f279e2ff3fbcea2b6e57ef1f4c58e01ad7040488a64ae02111aa4318332c6b not found: ID does not exist" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.946661 4993 scope.go:117] "RemoveContainer" containerID="d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6" Jan 29 12:32:59 crc kubenswrapper[4993]: E0129 12:32:59.946881 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6\": container with ID starting with d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6 not found: ID does not exist" containerID="d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6" Jan 29 12:32:59 crc kubenswrapper[4993]: I0129 12:32:59.946909 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6"} err="failed to get container status \"d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6\": rpc error: code = NotFound desc = could not find container \"d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6\": container with ID starting with d5f754f108c8e3b3ddff63d363ff063f1cd473ac7fa3d0d56ab2c23ea89514f6 not found: ID does not exist" Jan 29 12:33:01 crc kubenswrapper[4993]: I0129 12:33:01.199915 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" path="/var/lib/kubelet/pods/ba0a684d-6f4e-4c1f-8480-6d7167750711/volumes" Jan 29 12:33:05 crc kubenswrapper[4993]: I0129 12:33:05.190293 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:33:05 crc kubenswrapper[4993]: E0129 12:33:05.190770 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:33:18 crc kubenswrapper[4993]: I0129 12:33:18.191172 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:33:18 crc kubenswrapper[4993]: E0129 12:33:18.191942 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:33:33 crc kubenswrapper[4993]: I0129 12:33:33.190719 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:33:33 crc kubenswrapper[4993]: E0129 12:33:33.192785 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:33:48 crc kubenswrapper[4993]: I0129 12:33:48.191379 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:33:48 crc kubenswrapper[4993]: E0129 12:33:48.192074 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:34:03 crc kubenswrapper[4993]: I0129 12:34:03.190086 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:34:03 crc kubenswrapper[4993]: E0129 12:34:03.190851 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:34:15 crc kubenswrapper[4993]: I0129 12:34:15.191002 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:34:15 crc kubenswrapper[4993]: E0129 12:34:15.192086 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:34:26 crc kubenswrapper[4993]: I0129 12:34:26.190342 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:34:26 crc kubenswrapper[4993]: E0129 12:34:26.192023 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:34:38 crc kubenswrapper[4993]: I0129 12:34:38.190556 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:34:38 crc kubenswrapper[4993]: E0129 12:34:38.191697 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:34:52 crc kubenswrapper[4993]: I0129 12:34:52.191003 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:34:52 crc kubenswrapper[4993]: E0129 12:34:52.193883 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:35:03 crc kubenswrapper[4993]: I0129 12:35:03.190943 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:35:03 crc kubenswrapper[4993]: E0129 12:35:03.192409 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:35:15 crc kubenswrapper[4993]: I0129 12:35:15.190860 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:35:15 crc kubenswrapper[4993]: E0129 12:35:15.191608 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:35:30 crc kubenswrapper[4993]: I0129 12:35:30.191177 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:35:30 crc kubenswrapper[4993]: I0129 12:35:30.946696 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974"} Jan 29 12:37:48 crc kubenswrapper[4993]: I0129 12:37:48.156306 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:37:48 crc kubenswrapper[4993]: I0129 12:37:48.157169 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:38:18 crc kubenswrapper[4993]: I0129 12:38:18.156486 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:38:18 crc kubenswrapper[4993]: I0129 12:38:18.157129 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.155747 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.156462 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.156526 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.157269 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.157330 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974" gracePeriod=600 Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.463258 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974" exitCode=0 Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.463340 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974"} Jan 29 12:38:48 crc kubenswrapper[4993]: I0129 12:38:48.463858 4993 scope.go:117] "RemoveContainer" containerID="5b1e1a9ac92399d7d896596f0c8a390b54679a1456bac83eff50f189487102c1" Jan 29 12:38:49 crc kubenswrapper[4993]: I0129 12:38:49.474633 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779"} Jan 29 12:40:48 crc kubenswrapper[4993]: I0129 12:40:48.155584 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:40:48 crc kubenswrapper[4993]: I0129 12:40:48.156114 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.988142 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:04 crc kubenswrapper[4993]: E0129 12:41:04.989038 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="extract-utilities" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.989057 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="extract-utilities" Jan 29 12:41:04 crc kubenswrapper[4993]: E0129 12:41:04.989081 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="extract-content" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.989089 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="extract-content" Jan 29 12:41:04 crc kubenswrapper[4993]: E0129 12:41:04.989103 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="registry-server" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.989110 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="registry-server" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.989285 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba0a684d-6f4e-4c1f-8480-6d7167750711" containerName="registry-server" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.990296 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.994946 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9ndn\" (UniqueName: \"kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.994988 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:04 crc kubenswrapper[4993]: I0129 12:41:04.995033 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.017594 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.097385 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9ndn\" (UniqueName: \"kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.097466 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.097544 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.098041 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.098123 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.118119 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9ndn\" (UniqueName: \"kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn\") pod \"redhat-operators-9hdfx\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.317965 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:05 crc kubenswrapper[4993]: I0129 12:41:05.743093 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:05 crc kubenswrapper[4993]: W0129 12:41:05.748392 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe661ee9_362e_4a46_aac7_a5cff2f3ec69.slice/crio-2e251f5899565d3f6b2c64820c9865bd45df3096ca4a2af6ee7b82befe0f261e WatchSource:0}: Error finding container 2e251f5899565d3f6b2c64820c9865bd45df3096ca4a2af6ee7b82befe0f261e: Status 404 returned error can't find the container with id 2e251f5899565d3f6b2c64820c9865bd45df3096ca4a2af6ee7b82befe0f261e Jan 29 12:41:06 crc kubenswrapper[4993]: I0129 12:41:06.435700 4993 generic.go:334] "Generic (PLEG): container finished" podID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerID="04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595" exitCode=0 Jan 29 12:41:06 crc kubenswrapper[4993]: I0129 12:41:06.435757 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerDied","Data":"04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595"} Jan 29 12:41:06 crc kubenswrapper[4993]: I0129 12:41:06.435962 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerStarted","Data":"2e251f5899565d3f6b2c64820c9865bd45df3096ca4a2af6ee7b82befe0f261e"} Jan 29 12:41:06 crc kubenswrapper[4993]: I0129 12:41:06.436972 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:41:07 crc kubenswrapper[4993]: I0129 12:41:07.444828 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerStarted","Data":"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6"} Jan 29 12:41:08 crc kubenswrapper[4993]: I0129 12:41:08.454514 4993 generic.go:334] "Generic (PLEG): container finished" podID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerID="1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6" exitCode=0 Jan 29 12:41:08 crc kubenswrapper[4993]: I0129 12:41:08.454559 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerDied","Data":"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6"} Jan 29 12:41:09 crc kubenswrapper[4993]: I0129 12:41:09.464691 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerStarted","Data":"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a"} Jan 29 12:41:09 crc kubenswrapper[4993]: I0129 12:41:09.485295 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9hdfx" podStartSLOduration=3.082343936 podStartE2EDuration="5.485276394s" podCreationTimestamp="2026-01-29 12:41:04 +0000 UTC" firstStartedPulling="2026-01-29 12:41:06.43678617 +0000 UTC m=+2150.451916296" lastFinishedPulling="2026-01-29 12:41:08.839718628 +0000 UTC m=+2152.854848754" observedRunningTime="2026-01-29 12:41:09.479848907 +0000 UTC m=+2153.494979043" watchObservedRunningTime="2026-01-29 12:41:09.485276394 +0000 UTC m=+2153.500406520" Jan 29 12:41:15 crc kubenswrapper[4993]: I0129 12:41:15.318374 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:15 crc kubenswrapper[4993]: I0129 12:41:15.318781 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:15 crc kubenswrapper[4993]: I0129 12:41:15.367750 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:15 crc kubenswrapper[4993]: I0129 12:41:15.572672 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:15 crc kubenswrapper[4993]: I0129 12:41:15.631830 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:17 crc kubenswrapper[4993]: I0129 12:41:17.524565 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9hdfx" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="registry-server" containerID="cri-o://2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a" gracePeriod=2 Jan 29 12:41:17 crc kubenswrapper[4993]: I0129 12:41:17.935988 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.083423 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content\") pod \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.083476 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities\") pod \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.083540 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9ndn\" (UniqueName: \"kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn\") pod \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\" (UID: \"fe661ee9-362e-4a46-aac7-a5cff2f3ec69\") " Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.085028 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities" (OuterVolumeSpecName: "utilities") pod "fe661ee9-362e-4a46-aac7-a5cff2f3ec69" (UID: "fe661ee9-362e-4a46-aac7-a5cff2f3ec69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.089101 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn" (OuterVolumeSpecName: "kube-api-access-c9ndn") pod "fe661ee9-362e-4a46-aac7-a5cff2f3ec69" (UID: "fe661ee9-362e-4a46-aac7-a5cff2f3ec69"). InnerVolumeSpecName "kube-api-access-c9ndn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.155639 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.155694 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.184990 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.185021 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9ndn\" (UniqueName: \"kubernetes.io/projected/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-kube-api-access-c9ndn\") on node \"crc\" DevicePath \"\"" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.199489 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe661ee9-362e-4a46-aac7-a5cff2f3ec69" (UID: "fe661ee9-362e-4a46-aac7-a5cff2f3ec69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.286207 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe661ee9-362e-4a46-aac7-a5cff2f3ec69-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.537319 4993 generic.go:334] "Generic (PLEG): container finished" podID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerID="2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a" exitCode=0 Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.537362 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerDied","Data":"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a"} Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.537390 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9hdfx" event={"ID":"fe661ee9-362e-4a46-aac7-a5cff2f3ec69","Type":"ContainerDied","Data":"2e251f5899565d3f6b2c64820c9865bd45df3096ca4a2af6ee7b82befe0f261e"} Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.537407 4993 scope.go:117] "RemoveContainer" containerID="2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.537419 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9hdfx" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.559879 4993 scope.go:117] "RemoveContainer" containerID="1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.580927 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.586879 4993 scope.go:117] "RemoveContainer" containerID="04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.589080 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9hdfx"] Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.612263 4993 scope.go:117] "RemoveContainer" containerID="2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a" Jan 29 12:41:18 crc kubenswrapper[4993]: E0129 12:41:18.613180 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a\": container with ID starting with 2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a not found: ID does not exist" containerID="2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.613234 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a"} err="failed to get container status \"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a\": rpc error: code = NotFound desc = could not find container \"2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a\": container with ID starting with 2f0772333e8429cd904958f9f952122e88f1364fe1604a947ea8b540ca3b236a not found: ID does not exist" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.613257 4993 scope.go:117] "RemoveContainer" containerID="1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6" Jan 29 12:41:18 crc kubenswrapper[4993]: E0129 12:41:18.613602 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6\": container with ID starting with 1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6 not found: ID does not exist" containerID="1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.613619 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6"} err="failed to get container status \"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6\": rpc error: code = NotFound desc = could not find container \"1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6\": container with ID starting with 1a31bf6ece45202020e802f9c28491811adb7b4cada67ba3fb23be6131a680e6 not found: ID does not exist" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.613631 4993 scope.go:117] "RemoveContainer" containerID="04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595" Jan 29 12:41:18 crc kubenswrapper[4993]: E0129 12:41:18.613912 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595\": container with ID starting with 04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595 not found: ID does not exist" containerID="04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595" Jan 29 12:41:18 crc kubenswrapper[4993]: I0129 12:41:18.613938 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595"} err="failed to get container status \"04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595\": rpc error: code = NotFound desc = could not find container \"04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595\": container with ID starting with 04c6251b74383eefb0eedc5aa0679e21b3710a6675848902d21390f952215595 not found: ID does not exist" Jan 29 12:41:19 crc kubenswrapper[4993]: I0129 12:41:19.205516 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" path="/var/lib/kubelet/pods/fe661ee9-362e-4a46-aac7-a5cff2f3ec69/volumes" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.156447 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.157020 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.157075 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.157683 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.157758 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" gracePeriod=600 Jan 29 12:41:48 crc kubenswrapper[4993]: E0129 12:41:48.284692 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.770299 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" exitCode=0 Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.770372 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779"} Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.770446 4993 scope.go:117] "RemoveContainer" containerID="d4ac2a980848c32a822b5995a9c81058e048bece14c8c65ebd6929f4f8346974" Jan 29 12:41:48 crc kubenswrapper[4993]: I0129 12:41:48.772126 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:41:48 crc kubenswrapper[4993]: E0129 12:41:48.773000 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:42:01 crc kubenswrapper[4993]: I0129 12:42:01.197894 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:42:01 crc kubenswrapper[4993]: E0129 12:42:01.199133 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.078708 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:15 crc kubenswrapper[4993]: E0129 12:42:15.079436 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="extract-content" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.079447 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="extract-content" Jan 29 12:42:15 crc kubenswrapper[4993]: E0129 12:42:15.079461 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="extract-utilities" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.079468 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="extract-utilities" Jan 29 12:42:15 crc kubenswrapper[4993]: E0129 12:42:15.079484 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="registry-server" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.079490 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="registry-server" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.079630 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe661ee9-362e-4a46-aac7-a5cff2f3ec69" containerName="registry-server" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.087341 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.090669 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.190221 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:42:15 crc kubenswrapper[4993]: E0129 12:42:15.190450 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.257771 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.257822 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtcfz\" (UniqueName: \"kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.257851 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.358789 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.358912 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtcfz\" (UniqueName: \"kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.358946 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.359371 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.359789 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.382509 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtcfz\" (UniqueName: \"kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz\") pod \"redhat-marketplace-l8f2l\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.413325 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:15 crc kubenswrapper[4993]: I0129 12:42:15.863798 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:16 crc kubenswrapper[4993]: I0129 12:42:16.011475 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerStarted","Data":"2d4911825b44e94a6b3d6a3bdcfb27cd9b81fbab1193dcb786b37d5d1027394c"} Jan 29 12:42:17 crc kubenswrapper[4993]: I0129 12:42:17.020060 4993 generic.go:334] "Generic (PLEG): container finished" podID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerID="6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e" exitCode=0 Jan 29 12:42:17 crc kubenswrapper[4993]: I0129 12:42:17.020132 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerDied","Data":"6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e"} Jan 29 12:42:18 crc kubenswrapper[4993]: I0129 12:42:18.032666 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerStarted","Data":"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe"} Jan 29 12:42:19 crc kubenswrapper[4993]: I0129 12:42:19.049077 4993 generic.go:334] "Generic (PLEG): container finished" podID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerID="4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe" exitCode=0 Jan 29 12:42:19 crc kubenswrapper[4993]: I0129 12:42:19.049132 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerDied","Data":"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe"} Jan 29 12:42:20 crc kubenswrapper[4993]: I0129 12:42:20.058544 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerStarted","Data":"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152"} Jan 29 12:42:20 crc kubenswrapper[4993]: I0129 12:42:20.082146 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l8f2l" podStartSLOduration=2.619482438 podStartE2EDuration="5.082129727s" podCreationTimestamp="2026-01-29 12:42:15 +0000 UTC" firstStartedPulling="2026-01-29 12:42:17.024608018 +0000 UTC m=+2221.039738184" lastFinishedPulling="2026-01-29 12:42:19.487255307 +0000 UTC m=+2223.502385473" observedRunningTime="2026-01-29 12:42:20.075390174 +0000 UTC m=+2224.090520320" watchObservedRunningTime="2026-01-29 12:42:20.082129727 +0000 UTC m=+2224.097259853" Jan 29 12:42:25 crc kubenswrapper[4993]: I0129 12:42:25.413594 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:25 crc kubenswrapper[4993]: I0129 12:42:25.413958 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:25 crc kubenswrapper[4993]: I0129 12:42:25.470090 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:26 crc kubenswrapper[4993]: I0129 12:42:26.177713 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:26 crc kubenswrapper[4993]: I0129 12:42:26.227842 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.137501 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l8f2l" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="registry-server" containerID="cri-o://4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152" gracePeriod=2 Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.537682 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.689130 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtcfz\" (UniqueName: \"kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz\") pod \"eda6ccec-beb0-490c-8729-4c91cfe40c33\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.689201 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities\") pod \"eda6ccec-beb0-490c-8729-4c91cfe40c33\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.689288 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content\") pod \"eda6ccec-beb0-490c-8729-4c91cfe40c33\" (UID: \"eda6ccec-beb0-490c-8729-4c91cfe40c33\") " Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.690129 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities" (OuterVolumeSpecName: "utilities") pod "eda6ccec-beb0-490c-8729-4c91cfe40c33" (UID: "eda6ccec-beb0-490c-8729-4c91cfe40c33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.694436 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz" (OuterVolumeSpecName: "kube-api-access-mtcfz") pod "eda6ccec-beb0-490c-8729-4c91cfe40c33" (UID: "eda6ccec-beb0-490c-8729-4c91cfe40c33"). InnerVolumeSpecName "kube-api-access-mtcfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.710017 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eda6ccec-beb0-490c-8729-4c91cfe40c33" (UID: "eda6ccec-beb0-490c-8729-4c91cfe40c33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.791148 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtcfz\" (UniqueName: \"kubernetes.io/projected/eda6ccec-beb0-490c-8729-4c91cfe40c33-kube-api-access-mtcfz\") on node \"crc\" DevicePath \"\"" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.791206 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:42:28 crc kubenswrapper[4993]: I0129 12:42:28.791221 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eda6ccec-beb0-490c-8729-4c91cfe40c33-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.145692 4993 generic.go:334] "Generic (PLEG): container finished" podID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerID="4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152" exitCode=0 Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.145740 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerDied","Data":"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152"} Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.145766 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l8f2l" event={"ID":"eda6ccec-beb0-490c-8729-4c91cfe40c33","Type":"ContainerDied","Data":"2d4911825b44e94a6b3d6a3bdcfb27cd9b81fbab1193dcb786b37d5d1027394c"} Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.145787 4993 scope.go:117] "RemoveContainer" containerID="4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.145901 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l8f2l" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.174471 4993 scope.go:117] "RemoveContainer" containerID="4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.187321 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.203414 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l8f2l"] Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.204865 4993 scope.go:117] "RemoveContainer" containerID="6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.224376 4993 scope.go:117] "RemoveContainer" containerID="4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152" Jan 29 12:42:29 crc kubenswrapper[4993]: E0129 12:42:29.224731 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152\": container with ID starting with 4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152 not found: ID does not exist" containerID="4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.224776 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152"} err="failed to get container status \"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152\": rpc error: code = NotFound desc = could not find container \"4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152\": container with ID starting with 4dd5e92c0b1c7e109bdc224e73c11b52c6ba9a6d714667a1c86b4858964be152 not found: ID does not exist" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.224794 4993 scope.go:117] "RemoveContainer" containerID="4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe" Jan 29 12:42:29 crc kubenswrapper[4993]: E0129 12:42:29.224984 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe\": container with ID starting with 4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe not found: ID does not exist" containerID="4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.225008 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe"} err="failed to get container status \"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe\": rpc error: code = NotFound desc = could not find container \"4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe\": container with ID starting with 4d923444491761b5d0d04b71d7871455c87271463f597fc63a460af8e6991bfe not found: ID does not exist" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.225024 4993 scope.go:117] "RemoveContainer" containerID="6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e" Jan 29 12:42:29 crc kubenswrapper[4993]: E0129 12:42:29.225243 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e\": container with ID starting with 6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e not found: ID does not exist" containerID="6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e" Jan 29 12:42:29 crc kubenswrapper[4993]: I0129 12:42:29.225263 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e"} err="failed to get container status \"6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e\": rpc error: code = NotFound desc = could not find container \"6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e\": container with ID starting with 6c557fb0e0745ace3c9cf85d2901a5455a65aa8b09733beb76efbb704c34a46e not found: ID does not exist" Jan 29 12:42:30 crc kubenswrapper[4993]: I0129 12:42:30.190601 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:42:30 crc kubenswrapper[4993]: E0129 12:42:30.190834 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:42:31 crc kubenswrapper[4993]: I0129 12:42:31.200271 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" path="/var/lib/kubelet/pods/eda6ccec-beb0-490c-8729-4c91cfe40c33/volumes" Jan 29 12:42:42 crc kubenswrapper[4993]: I0129 12:42:42.190646 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:42:42 crc kubenswrapper[4993]: E0129 12:42:42.191521 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:42:57 crc kubenswrapper[4993]: I0129 12:42:57.193718 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:42:57 crc kubenswrapper[4993]: E0129 12:42:57.194419 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.871149 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:00 crc kubenswrapper[4993]: E0129 12:43:00.871851 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="extract-content" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.871869 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="extract-content" Jan 29 12:43:00 crc kubenswrapper[4993]: E0129 12:43:00.871880 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="extract-utilities" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.871889 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="extract-utilities" Jan 29 12:43:00 crc kubenswrapper[4993]: E0129 12:43:00.871902 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="registry-server" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.871911 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="registry-server" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.872082 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda6ccec-beb0-490c-8729-4c91cfe40c33" containerName="registry-server" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.874486 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:00 crc kubenswrapper[4993]: I0129 12:43:00.888279 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.030475 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.030566 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf257\" (UniqueName: \"kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.030673 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.132474 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.132534 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf257\" (UniqueName: \"kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.132608 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.133088 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.133421 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.155904 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf257\" (UniqueName: \"kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257\") pod \"community-operators-s59dd\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.205094 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:01 crc kubenswrapper[4993]: I0129 12:43:01.698120 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:02 crc kubenswrapper[4993]: I0129 12:43:02.375412 4993 generic.go:334] "Generic (PLEG): container finished" podID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerID="a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529" exitCode=0 Jan 29 12:43:02 crc kubenswrapper[4993]: I0129 12:43:02.375475 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerDied","Data":"a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529"} Jan 29 12:43:02 crc kubenswrapper[4993]: I0129 12:43:02.375630 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerStarted","Data":"c7bb840e025dc500913c9777cf5bab8181f8e7e50814ee724e69d0d323b2d92a"} Jan 29 12:43:04 crc kubenswrapper[4993]: I0129 12:43:04.391875 4993 generic.go:334] "Generic (PLEG): container finished" podID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerID="a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791" exitCode=0 Jan 29 12:43:04 crc kubenswrapper[4993]: I0129 12:43:04.391921 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerDied","Data":"a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791"} Jan 29 12:43:06 crc kubenswrapper[4993]: I0129 12:43:06.405077 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerStarted","Data":"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b"} Jan 29 12:43:06 crc kubenswrapper[4993]: I0129 12:43:06.422907 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s59dd" podStartSLOduration=3.620508704 podStartE2EDuration="6.422889556s" podCreationTimestamp="2026-01-29 12:43:00 +0000 UTC" firstStartedPulling="2026-01-29 12:43:02.377107647 +0000 UTC m=+2266.392237773" lastFinishedPulling="2026-01-29 12:43:05.179488499 +0000 UTC m=+2269.194618625" observedRunningTime="2026-01-29 12:43:06.421811516 +0000 UTC m=+2270.436941642" watchObservedRunningTime="2026-01-29 12:43:06.422889556 +0000 UTC m=+2270.438019682" Jan 29 12:43:10 crc kubenswrapper[4993]: I0129 12:43:10.191036 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:43:10 crc kubenswrapper[4993]: E0129 12:43:10.191581 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:43:11 crc kubenswrapper[4993]: I0129 12:43:11.206805 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:11 crc kubenswrapper[4993]: I0129 12:43:11.207166 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:11 crc kubenswrapper[4993]: I0129 12:43:11.266456 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:11 crc kubenswrapper[4993]: I0129 12:43:11.476950 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:11 crc kubenswrapper[4993]: I0129 12:43:11.522955 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:13 crc kubenswrapper[4993]: I0129 12:43:13.465661 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s59dd" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="registry-server" containerID="cri-o://46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b" gracePeriod=2 Jan 29 12:43:13 crc kubenswrapper[4993]: I0129 12:43:13.886036 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.018241 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content\") pod \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.018317 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf257\" (UniqueName: \"kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257\") pod \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.018341 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities\") pod \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\" (UID: \"03e0d3ee-fff2-42a2-865e-b6befab6d5e1\") " Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.019798 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities" (OuterVolumeSpecName: "utilities") pod "03e0d3ee-fff2-42a2-865e-b6befab6d5e1" (UID: "03e0d3ee-fff2-42a2-865e-b6befab6d5e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.024477 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257" (OuterVolumeSpecName: "kube-api-access-gf257") pod "03e0d3ee-fff2-42a2-865e-b6befab6d5e1" (UID: "03e0d3ee-fff2-42a2-865e-b6befab6d5e1"). InnerVolumeSpecName "kube-api-access-gf257". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.082041 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03e0d3ee-fff2-42a2-865e-b6befab6d5e1" (UID: "03e0d3ee-fff2-42a2-865e-b6befab6d5e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.119651 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.120003 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf257\" (UniqueName: \"kubernetes.io/projected/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-kube-api-access-gf257\") on node \"crc\" DevicePath \"\"" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.120020 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e0d3ee-fff2-42a2-865e-b6befab6d5e1-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.478006 4993 generic.go:334] "Generic (PLEG): container finished" podID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerID="46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b" exitCode=0 Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.478051 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerDied","Data":"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b"} Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.478079 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s59dd" event={"ID":"03e0d3ee-fff2-42a2-865e-b6befab6d5e1","Type":"ContainerDied","Data":"c7bb840e025dc500913c9777cf5bab8181f8e7e50814ee724e69d0d323b2d92a"} Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.478096 4993 scope.go:117] "RemoveContainer" containerID="46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.478115 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s59dd" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.499030 4993 scope.go:117] "RemoveContainer" containerID="a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.514293 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.519354 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s59dd"] Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.539218 4993 scope.go:117] "RemoveContainer" containerID="a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.558928 4993 scope.go:117] "RemoveContainer" containerID="46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b" Jan 29 12:43:14 crc kubenswrapper[4993]: E0129 12:43:14.559425 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b\": container with ID starting with 46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b not found: ID does not exist" containerID="46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.559463 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b"} err="failed to get container status \"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b\": rpc error: code = NotFound desc = could not find container \"46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b\": container with ID starting with 46e5820a6d76b727d97e9cf0ccf7d356ab8b5cf565850ef462069c4832a7f37b not found: ID does not exist" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.559486 4993 scope.go:117] "RemoveContainer" containerID="a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791" Jan 29 12:43:14 crc kubenswrapper[4993]: E0129 12:43:14.559696 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791\": container with ID starting with a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791 not found: ID does not exist" containerID="a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.559716 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791"} err="failed to get container status \"a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791\": rpc error: code = NotFound desc = could not find container \"a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791\": container with ID starting with a27fa889d8ea3abc0b2c5b2ef939d4857b2c3d0fc6487f1821a701bdf916b791 not found: ID does not exist" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.559731 4993 scope.go:117] "RemoveContainer" containerID="a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529" Jan 29 12:43:14 crc kubenswrapper[4993]: E0129 12:43:14.559926 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529\": container with ID starting with a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529 not found: ID does not exist" containerID="a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529" Jan 29 12:43:14 crc kubenswrapper[4993]: I0129 12:43:14.559951 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529"} err="failed to get container status \"a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529\": rpc error: code = NotFound desc = could not find container \"a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529\": container with ID starting with a64e40b2145aad7346fbf565876d47bc6f06ddc27296393362626b17a0a82529 not found: ID does not exist" Jan 29 12:43:15 crc kubenswrapper[4993]: I0129 12:43:15.199430 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" path="/var/lib/kubelet/pods/03e0d3ee-fff2-42a2-865e-b6befab6d5e1/volumes" Jan 29 12:43:25 crc kubenswrapper[4993]: I0129 12:43:25.189912 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:43:25 crc kubenswrapper[4993]: E0129 12:43:25.190327 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:43:37 crc kubenswrapper[4993]: I0129 12:43:37.197181 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:43:37 crc kubenswrapper[4993]: E0129 12:43:37.198842 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:43:48 crc kubenswrapper[4993]: I0129 12:43:48.190546 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:43:48 crc kubenswrapper[4993]: E0129 12:43:48.191363 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:44:00 crc kubenswrapper[4993]: I0129 12:44:00.191165 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:44:00 crc kubenswrapper[4993]: E0129 12:44:00.192318 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:44:12 crc kubenswrapper[4993]: I0129 12:44:12.190505 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:44:12 crc kubenswrapper[4993]: E0129 12:44:12.191523 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:44:27 crc kubenswrapper[4993]: I0129 12:44:27.193675 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:44:27 crc kubenswrapper[4993]: E0129 12:44:27.194468 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:44:40 crc kubenswrapper[4993]: I0129 12:44:40.190227 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:44:40 crc kubenswrapper[4993]: E0129 12:44:40.191047 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:44:55 crc kubenswrapper[4993]: I0129 12:44:55.190680 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:44:55 crc kubenswrapper[4993]: E0129 12:44:55.191707 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.155646 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk"] Jan 29 12:45:00 crc kubenswrapper[4993]: E0129 12:45:00.156553 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="extract-utilities" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.156570 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="extract-utilities" Jan 29 12:45:00 crc kubenswrapper[4993]: E0129 12:45:00.156582 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="registry-server" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.156590 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="registry-server" Jan 29 12:45:00 crc kubenswrapper[4993]: E0129 12:45:00.156604 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="extract-content" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.156612 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="extract-content" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.156787 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="03e0d3ee-fff2-42a2-865e-b6befab6d5e1" containerName="registry-server" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.157325 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.160116 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.160155 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.171948 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk"] Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.264401 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hgn\" (UniqueName: \"kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.265330 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.265619 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.367414 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hgn\" (UniqueName: \"kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.367500 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.367573 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.368569 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.375656 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.385032 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hgn\" (UniqueName: \"kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn\") pod \"collect-profiles-29494845-xlbqk\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.481008 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:00 crc kubenswrapper[4993]: I0129 12:45:00.930674 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk"] Jan 29 12:45:01 crc kubenswrapper[4993]: I0129 12:45:01.343505 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" event={"ID":"6a591e34-1923-4aec-9273-8ffc6a72616e","Type":"ContainerStarted","Data":"4caceaa04a31aa685e90e2525a4d9245c759972474fc6b212adbdf675721b2d0"} Jan 29 12:45:01 crc kubenswrapper[4993]: I0129 12:45:01.343926 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" event={"ID":"6a591e34-1923-4aec-9273-8ffc6a72616e","Type":"ContainerStarted","Data":"6c54858806bff6eef2aa68135694c35e6f335ceaa2dbcb336c1c519a7da1ba94"} Jan 29 12:45:01 crc kubenswrapper[4993]: I0129 12:45:01.363998 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" podStartSLOduration=1.363979821 podStartE2EDuration="1.363979821s" podCreationTimestamp="2026-01-29 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 12:45:01.357619739 +0000 UTC m=+2385.372749865" watchObservedRunningTime="2026-01-29 12:45:01.363979821 +0000 UTC m=+2385.379109947" Jan 29 12:45:02 crc kubenswrapper[4993]: I0129 12:45:02.352707 4993 generic.go:334] "Generic (PLEG): container finished" podID="6a591e34-1923-4aec-9273-8ffc6a72616e" containerID="4caceaa04a31aa685e90e2525a4d9245c759972474fc6b212adbdf675721b2d0" exitCode=0 Jan 29 12:45:02 crc kubenswrapper[4993]: I0129 12:45:02.352811 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" event={"ID":"6a591e34-1923-4aec-9273-8ffc6a72616e","Type":"ContainerDied","Data":"4caceaa04a31aa685e90e2525a4d9245c759972474fc6b212adbdf675721b2d0"} Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.677317 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.729128 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6hgn\" (UniqueName: \"kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn\") pod \"6a591e34-1923-4aec-9273-8ffc6a72616e\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.729239 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume\") pod \"6a591e34-1923-4aec-9273-8ffc6a72616e\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.729318 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume\") pod \"6a591e34-1923-4aec-9273-8ffc6a72616e\" (UID: \"6a591e34-1923-4aec-9273-8ffc6a72616e\") " Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.730115 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume" (OuterVolumeSpecName: "config-volume") pod "6a591e34-1923-4aec-9273-8ffc6a72616e" (UID: "6a591e34-1923-4aec-9273-8ffc6a72616e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.734754 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn" (OuterVolumeSpecName: "kube-api-access-n6hgn") pod "6a591e34-1923-4aec-9273-8ffc6a72616e" (UID: "6a591e34-1923-4aec-9273-8ffc6a72616e"). InnerVolumeSpecName "kube-api-access-n6hgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.734814 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6a591e34-1923-4aec-9273-8ffc6a72616e" (UID: "6a591e34-1923-4aec-9273-8ffc6a72616e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.830528 4993 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a591e34-1923-4aec-9273-8ffc6a72616e-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.830594 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a591e34-1923-4aec-9273-8ffc6a72616e-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 12:45:03 crc kubenswrapper[4993]: I0129 12:45:03.830608 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6hgn\" (UniqueName: \"kubernetes.io/projected/6a591e34-1923-4aec-9273-8ffc6a72616e-kube-api-access-n6hgn\") on node \"crc\" DevicePath \"\"" Jan 29 12:45:04 crc kubenswrapper[4993]: I0129 12:45:04.366238 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" event={"ID":"6a591e34-1923-4aec-9273-8ffc6a72616e","Type":"ContainerDied","Data":"6c54858806bff6eef2aa68135694c35e6f335ceaa2dbcb336c1c519a7da1ba94"} Jan 29 12:45:04 crc kubenswrapper[4993]: I0129 12:45:04.366294 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c54858806bff6eef2aa68135694c35e6f335ceaa2dbcb336c1c519a7da1ba94" Jan 29 12:45:04 crc kubenswrapper[4993]: I0129 12:45:04.366269 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494845-xlbqk" Jan 29 12:45:04 crc kubenswrapper[4993]: I0129 12:45:04.424125 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb"] Jan 29 12:45:04 crc kubenswrapper[4993]: I0129 12:45:04.430973 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494800-mpzgb"] Jan 29 12:45:05 crc kubenswrapper[4993]: I0129 12:45:05.199596 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8835a1b6-a97d-48c8-9dd3-c197250e4825" path="/var/lib/kubelet/pods/8835a1b6-a97d-48c8-9dd3-c197250e4825/volumes" Jan 29 12:45:10 crc kubenswrapper[4993]: I0129 12:45:10.190517 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:45:10 crc kubenswrapper[4993]: E0129 12:45:10.191000 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:45:18 crc kubenswrapper[4993]: I0129 12:45:18.213871 4993 scope.go:117] "RemoveContainer" containerID="ccfc6e9565d9325d95e7125b923c9969c9c76b944792c576c6e12306ce990dd4" Jan 29 12:45:21 crc kubenswrapper[4993]: I0129 12:45:21.190836 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:45:21 crc kubenswrapper[4993]: E0129 12:45:21.191435 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:45:34 crc kubenswrapper[4993]: I0129 12:45:34.190005 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:45:34 crc kubenswrapper[4993]: E0129 12:45:34.190764 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:45:45 crc kubenswrapper[4993]: I0129 12:45:45.190868 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:45:45 crc kubenswrapper[4993]: E0129 12:45:45.192503 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:45:58 crc kubenswrapper[4993]: I0129 12:45:58.189790 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:45:58 crc kubenswrapper[4993]: E0129 12:45:58.190516 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:46:09 crc kubenswrapper[4993]: I0129 12:46:09.190550 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:46:09 crc kubenswrapper[4993]: E0129 12:46:09.191620 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:46:24 crc kubenswrapper[4993]: I0129 12:46:24.190648 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:46:24 crc kubenswrapper[4993]: E0129 12:46:24.191680 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:46:37 crc kubenswrapper[4993]: I0129 12:46:37.196831 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:46:37 crc kubenswrapper[4993]: E0129 12:46:37.199063 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:46:51 crc kubenswrapper[4993]: I0129 12:46:51.190559 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:46:52 crc kubenswrapper[4993]: I0129 12:46:52.286620 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be"} Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.057654 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:41 crc kubenswrapper[4993]: E0129 12:47:41.060924 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a591e34-1923-4aec-9273-8ffc6a72616e" containerName="collect-profiles" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.060941 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a591e34-1923-4aec-9273-8ffc6a72616e" containerName="collect-profiles" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.061477 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a591e34-1923-4aec-9273-8ffc6a72616e" containerName="collect-profiles" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.062956 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.071789 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.072052 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.072253 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdpv9\" (UniqueName: \"kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.082234 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.174123 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.174207 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.174309 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdpv9\" (UniqueName: \"kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.174850 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.175034 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.199701 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdpv9\" (UniqueName: \"kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9\") pod \"certified-operators-j9s2x\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.406708 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:41 crc kubenswrapper[4993]: I0129 12:47:41.843929 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:42 crc kubenswrapper[4993]: I0129 12:47:42.716366 4993 generic.go:334] "Generic (PLEG): container finished" podID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerID="e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07" exitCode=0 Jan 29 12:47:42 crc kubenswrapper[4993]: I0129 12:47:42.716467 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerDied","Data":"e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07"} Jan 29 12:47:42 crc kubenswrapper[4993]: I0129 12:47:42.716633 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerStarted","Data":"fc86c91818b4d94de25194121a02217bb254253001107fdaee80c090db7f3900"} Jan 29 12:47:42 crc kubenswrapper[4993]: I0129 12:47:42.720207 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:47:44 crc kubenswrapper[4993]: I0129 12:47:44.742532 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerStarted","Data":"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6"} Jan 29 12:47:45 crc kubenswrapper[4993]: I0129 12:47:45.754384 4993 generic.go:334] "Generic (PLEG): container finished" podID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerID="503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6" exitCode=0 Jan 29 12:47:45 crc kubenswrapper[4993]: I0129 12:47:45.754450 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerDied","Data":"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6"} Jan 29 12:47:47 crc kubenswrapper[4993]: I0129 12:47:47.781158 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerStarted","Data":"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3"} Jan 29 12:47:47 crc kubenswrapper[4993]: I0129 12:47:47.809043 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j9s2x" podStartSLOduration=2.936718793 podStartE2EDuration="6.809019375s" podCreationTimestamp="2026-01-29 12:47:41 +0000 UTC" firstStartedPulling="2026-01-29 12:47:42.719960025 +0000 UTC m=+2546.735090151" lastFinishedPulling="2026-01-29 12:47:46.592260597 +0000 UTC m=+2550.607390733" observedRunningTime="2026-01-29 12:47:47.804668407 +0000 UTC m=+2551.819798533" watchObservedRunningTime="2026-01-29 12:47:47.809019375 +0000 UTC m=+2551.824149511" Jan 29 12:47:51 crc kubenswrapper[4993]: I0129 12:47:51.407274 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:51 crc kubenswrapper[4993]: I0129 12:47:51.407705 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:51 crc kubenswrapper[4993]: I0129 12:47:51.479845 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:51 crc kubenswrapper[4993]: I0129 12:47:51.861231 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:53 crc kubenswrapper[4993]: I0129 12:47:53.075809 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:53 crc kubenswrapper[4993]: I0129 12:47:53.825763 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j9s2x" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="registry-server" containerID="cri-o://e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3" gracePeriod=2 Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.271604 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.377218 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content\") pod \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.377281 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdpv9\" (UniqueName: \"kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9\") pod \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.377351 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities\") pod \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\" (UID: \"2f2e0a94-e07e-4d5f-adba-1b704bdb3288\") " Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.378405 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities" (OuterVolumeSpecName: "utilities") pod "2f2e0a94-e07e-4d5f-adba-1b704bdb3288" (UID: "2f2e0a94-e07e-4d5f-adba-1b704bdb3288"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.382240 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9" (OuterVolumeSpecName: "kube-api-access-fdpv9") pod "2f2e0a94-e07e-4d5f-adba-1b704bdb3288" (UID: "2f2e0a94-e07e-4d5f-adba-1b704bdb3288"). InnerVolumeSpecName "kube-api-access-fdpv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.431982 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f2e0a94-e07e-4d5f-adba-1b704bdb3288" (UID: "2f2e0a94-e07e-4d5f-adba-1b704bdb3288"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.478989 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.479032 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdpv9\" (UniqueName: \"kubernetes.io/projected/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-kube-api-access-fdpv9\") on node \"crc\" DevicePath \"\"" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.479050 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2e0a94-e07e-4d5f-adba-1b704bdb3288-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.836776 4993 generic.go:334] "Generic (PLEG): container finished" podID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerID="e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3" exitCode=0 Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.836825 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerDied","Data":"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3"} Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.836853 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j9s2x" event={"ID":"2f2e0a94-e07e-4d5f-adba-1b704bdb3288","Type":"ContainerDied","Data":"fc86c91818b4d94de25194121a02217bb254253001107fdaee80c090db7f3900"} Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.836870 4993 scope.go:117] "RemoveContainer" containerID="e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.836875 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j9s2x" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.861460 4993 scope.go:117] "RemoveContainer" containerID="503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.876581 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.882881 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j9s2x"] Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.897261 4993 scope.go:117] "RemoveContainer" containerID="e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.913885 4993 scope.go:117] "RemoveContainer" containerID="e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3" Jan 29 12:47:54 crc kubenswrapper[4993]: E0129 12:47:54.914351 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3\": container with ID starting with e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3 not found: ID does not exist" containerID="e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.914384 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3"} err="failed to get container status \"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3\": rpc error: code = NotFound desc = could not find container \"e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3\": container with ID starting with e236c79852bc6dbc53c28949eb03f80fb49ec66b6d082461bd94571b067b91c3 not found: ID does not exist" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.914402 4993 scope.go:117] "RemoveContainer" containerID="503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6" Jan 29 12:47:54 crc kubenswrapper[4993]: E0129 12:47:54.914700 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6\": container with ID starting with 503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6 not found: ID does not exist" containerID="503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.914730 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6"} err="failed to get container status \"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6\": rpc error: code = NotFound desc = could not find container \"503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6\": container with ID starting with 503e1c8df56ac48b7b8f6698f8b34164ea60cba89a43215a57aac6ed0967c8f6 not found: ID does not exist" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.914748 4993 scope.go:117] "RemoveContainer" containerID="e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07" Jan 29 12:47:54 crc kubenswrapper[4993]: E0129 12:47:54.915004 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07\": container with ID starting with e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07 not found: ID does not exist" containerID="e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07" Jan 29 12:47:54 crc kubenswrapper[4993]: I0129 12:47:54.915031 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07"} err="failed to get container status \"e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07\": rpc error: code = NotFound desc = could not find container \"e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07\": container with ID starting with e601de9044b6cc991b6e7367ccf4d99e18da36a060036ed7cdb99d142bf08c07 not found: ID does not exist" Jan 29 12:47:55 crc kubenswrapper[4993]: I0129 12:47:55.199138 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" path="/var/lib/kubelet/pods/2f2e0a94-e07e-4d5f-adba-1b704bdb3288/volumes" Jan 29 12:49:18 crc kubenswrapper[4993]: I0129 12:49:18.156393 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:49:18 crc kubenswrapper[4993]: I0129 12:49:18.157569 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:49:48 crc kubenswrapper[4993]: I0129 12:49:48.156238 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:49:48 crc kubenswrapper[4993]: I0129 12:49:48.157103 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:50:18 crc kubenswrapper[4993]: I0129 12:50:18.156345 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:50:18 crc kubenswrapper[4993]: I0129 12:50:18.156892 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:50:18 crc kubenswrapper[4993]: I0129 12:50:18.156936 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:50:18 crc kubenswrapper[4993]: I0129 12:50:18.157524 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:50:18 crc kubenswrapper[4993]: I0129 12:50:18.157577 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be" gracePeriod=600 Jan 29 12:50:19 crc kubenswrapper[4993]: I0129 12:50:19.282562 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be" exitCode=0 Jan 29 12:50:19 crc kubenswrapper[4993]: I0129 12:50:19.282655 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be"} Jan 29 12:50:19 crc kubenswrapper[4993]: I0129 12:50:19.284047 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f"} Jan 29 12:50:19 crc kubenswrapper[4993]: I0129 12:50:19.284091 4993 scope.go:117] "RemoveContainer" containerID="69eed9373f0dfb63aa638ed6f0d0df8f2d01199a4cd6640f0cb28d1e52710779" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.442036 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:18 crc kubenswrapper[4993]: E0129 12:51:18.443018 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="extract-utilities" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.443039 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="extract-utilities" Jan 29 12:51:18 crc kubenswrapper[4993]: E0129 12:51:18.443057 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="registry-server" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.443066 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="registry-server" Jan 29 12:51:18 crc kubenswrapper[4993]: E0129 12:51:18.443080 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="extract-content" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.443089 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="extract-content" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.443295 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f2e0a94-e07e-4d5f-adba-1b704bdb3288" containerName="registry-server" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.444924 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.448802 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.629036 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.629176 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.629364 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fnqn\" (UniqueName: \"kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.730302 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.730375 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fnqn\" (UniqueName: \"kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.730408 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.730883 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.730916 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.753383 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fnqn\" (UniqueName: \"kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn\") pod \"redhat-operators-stzfs\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.764712 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:18 crc kubenswrapper[4993]: I0129 12:51:18.992310 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:19 crc kubenswrapper[4993]: I0129 12:51:19.742902 4993 generic.go:334] "Generic (PLEG): container finished" podID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerID="60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf" exitCode=0 Jan 29 12:51:19 crc kubenswrapper[4993]: I0129 12:51:19.743025 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerDied","Data":"60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf"} Jan 29 12:51:19 crc kubenswrapper[4993]: I0129 12:51:19.743266 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerStarted","Data":"ffa490c32b36bef791aadc2d9db0013448c6f7169c552565f7138b01d136a2a1"} Jan 29 12:51:22 crc kubenswrapper[4993]: I0129 12:51:22.765920 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerStarted","Data":"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77"} Jan 29 12:51:23 crc kubenswrapper[4993]: I0129 12:51:23.774966 4993 generic.go:334] "Generic (PLEG): container finished" podID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerID="28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77" exitCode=0 Jan 29 12:51:23 crc kubenswrapper[4993]: I0129 12:51:23.775010 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerDied","Data":"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77"} Jan 29 12:51:25 crc kubenswrapper[4993]: I0129 12:51:25.791947 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerStarted","Data":"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227"} Jan 29 12:51:25 crc kubenswrapper[4993]: I0129 12:51:25.817525 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-stzfs" podStartSLOduration=2.486346078 podStartE2EDuration="7.8175072s" podCreationTimestamp="2026-01-29 12:51:18 +0000 UTC" firstStartedPulling="2026-01-29 12:51:19.745501097 +0000 UTC m=+2763.760631223" lastFinishedPulling="2026-01-29 12:51:25.076662229 +0000 UTC m=+2769.091792345" observedRunningTime="2026-01-29 12:51:25.814122689 +0000 UTC m=+2769.829252815" watchObservedRunningTime="2026-01-29 12:51:25.8175072 +0000 UTC m=+2769.832637326" Jan 29 12:51:28 crc kubenswrapper[4993]: I0129 12:51:28.766308 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:28 crc kubenswrapper[4993]: I0129 12:51:28.766544 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:29 crc kubenswrapper[4993]: I0129 12:51:29.819879 4993 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-stzfs" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="registry-server" probeResult="failure" output=< Jan 29 12:51:29 crc kubenswrapper[4993]: timeout: failed to connect service ":50051" within 1s Jan 29 12:51:29 crc kubenswrapper[4993]: > Jan 29 12:51:38 crc kubenswrapper[4993]: I0129 12:51:38.817741 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:38 crc kubenswrapper[4993]: I0129 12:51:38.858242 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:39 crc kubenswrapper[4993]: I0129 12:51:39.054409 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:39 crc kubenswrapper[4993]: I0129 12:51:39.884834 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-stzfs" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="registry-server" containerID="cri-o://4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227" gracePeriod=2 Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.282366 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.341448 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fnqn\" (UniqueName: \"kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn\") pod \"cb96fae2-5fff-4536-970d-b9ce97ee6784\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.341520 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content\") pod \"cb96fae2-5fff-4536-970d-b9ce97ee6784\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.341548 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities\") pod \"cb96fae2-5fff-4536-970d-b9ce97ee6784\" (UID: \"cb96fae2-5fff-4536-970d-b9ce97ee6784\") " Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.342395 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities" (OuterVolumeSpecName: "utilities") pod "cb96fae2-5fff-4536-970d-b9ce97ee6784" (UID: "cb96fae2-5fff-4536-970d-b9ce97ee6784"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.342554 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.347734 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn" (OuterVolumeSpecName: "kube-api-access-8fnqn") pod "cb96fae2-5fff-4536-970d-b9ce97ee6784" (UID: "cb96fae2-5fff-4536-970d-b9ce97ee6784"). InnerVolumeSpecName "kube-api-access-8fnqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.443168 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fnqn\" (UniqueName: \"kubernetes.io/projected/cb96fae2-5fff-4536-970d-b9ce97ee6784-kube-api-access-8fnqn\") on node \"crc\" DevicePath \"\"" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.452160 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb96fae2-5fff-4536-970d-b9ce97ee6784" (UID: "cb96fae2-5fff-4536-970d-b9ce97ee6784"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.544265 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb96fae2-5fff-4536-970d-b9ce97ee6784-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.894293 4993 generic.go:334] "Generic (PLEG): container finished" podID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerID="4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227" exitCode=0 Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.894346 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerDied","Data":"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227"} Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.894408 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stzfs" event={"ID":"cb96fae2-5fff-4536-970d-b9ce97ee6784","Type":"ContainerDied","Data":"ffa490c32b36bef791aadc2d9db0013448c6f7169c552565f7138b01d136a2a1"} Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.894414 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stzfs" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.894434 4993 scope.go:117] "RemoveContainer" containerID="4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.913042 4993 scope.go:117] "RemoveContainer" containerID="28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.933881 4993 scope.go:117] "RemoveContainer" containerID="60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.963465 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.969250 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-stzfs"] Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.973113 4993 scope.go:117] "RemoveContainer" containerID="4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227" Jan 29 12:51:40 crc kubenswrapper[4993]: E0129 12:51:40.973563 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227\": container with ID starting with 4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227 not found: ID does not exist" containerID="4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.973671 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227"} err="failed to get container status \"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227\": rpc error: code = NotFound desc = could not find container \"4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227\": container with ID starting with 4f10a0c96d77bbdadc8faa5d18ec740a251919d1bbe1fa64fc5661d7ae788227 not found: ID does not exist" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.973759 4993 scope.go:117] "RemoveContainer" containerID="28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77" Jan 29 12:51:40 crc kubenswrapper[4993]: E0129 12:51:40.974218 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77\": container with ID starting with 28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77 not found: ID does not exist" containerID="28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.974306 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77"} err="failed to get container status \"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77\": rpc error: code = NotFound desc = could not find container \"28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77\": container with ID starting with 28b223513a670ecd1bd03d53f42ec95dbb43123dda57d5d856684fe47b606e77 not found: ID does not exist" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.974384 4993 scope.go:117] "RemoveContainer" containerID="60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf" Jan 29 12:51:40 crc kubenswrapper[4993]: E0129 12:51:40.974730 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf\": container with ID starting with 60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf not found: ID does not exist" containerID="60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf" Jan 29 12:51:40 crc kubenswrapper[4993]: I0129 12:51:40.974764 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf"} err="failed to get container status \"60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf\": rpc error: code = NotFound desc = could not find container \"60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf\": container with ID starting with 60ebc75251cc281965cf0128358f02fb54826a87d8ea2e140812c491707ed1cf not found: ID does not exist" Jan 29 12:51:41 crc kubenswrapper[4993]: I0129 12:51:41.198809 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" path="/var/lib/kubelet/pods/cb96fae2-5fff-4536-970d-b9ce97ee6784/volumes" Jan 29 12:52:18 crc kubenswrapper[4993]: I0129 12:52:18.156369 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:52:18 crc kubenswrapper[4993]: I0129 12:52:18.157051 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.561110 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:31 crc kubenswrapper[4993]: E0129 12:52:31.562609 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="registry-server" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.562634 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="registry-server" Jan 29 12:52:31 crc kubenswrapper[4993]: E0129 12:52:31.562671 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="extract-content" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.562684 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="extract-content" Jan 29 12:52:31 crc kubenswrapper[4993]: E0129 12:52:31.562716 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="extract-utilities" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.562732 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="extract-utilities" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.563001 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb96fae2-5fff-4536-970d-b9ce97ee6784" containerName="registry-server" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.564878 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.572520 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.683267 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4r9c\" (UniqueName: \"kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.683575 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.683719 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.785211 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.785296 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4r9c\" (UniqueName: \"kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.785351 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.785827 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.785950 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.806814 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4r9c\" (UniqueName: \"kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c\") pod \"redhat-marketplace-hvg8l\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:31 crc kubenswrapper[4993]: I0129 12:52:31.905342 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:32 crc kubenswrapper[4993]: I0129 12:52:32.331296 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:33 crc kubenswrapper[4993]: I0129 12:52:33.323119 4993 generic.go:334] "Generic (PLEG): container finished" podID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerID="f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab" exitCode=0 Jan 29 12:52:33 crc kubenswrapper[4993]: I0129 12:52:33.323250 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerDied","Data":"f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab"} Jan 29 12:52:33 crc kubenswrapper[4993]: I0129 12:52:33.323650 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerStarted","Data":"ec6c92a76f5ae463f3db10414bbbecd96f843e6eb8b43b2f027280d097f51b63"} Jan 29 12:52:34 crc kubenswrapper[4993]: I0129 12:52:34.334381 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerStarted","Data":"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90"} Jan 29 12:52:35 crc kubenswrapper[4993]: I0129 12:52:35.346387 4993 generic.go:334] "Generic (PLEG): container finished" podID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerID="eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90" exitCode=0 Jan 29 12:52:35 crc kubenswrapper[4993]: I0129 12:52:35.346489 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerDied","Data":"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90"} Jan 29 12:52:36 crc kubenswrapper[4993]: I0129 12:52:36.354442 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerStarted","Data":"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971"} Jan 29 12:52:36 crc kubenswrapper[4993]: I0129 12:52:36.380309 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hvg8l" podStartSLOduration=2.804417754 podStartE2EDuration="5.38028732s" podCreationTimestamp="2026-01-29 12:52:31 +0000 UTC" firstStartedPulling="2026-01-29 12:52:33.325707415 +0000 UTC m=+2837.340837541" lastFinishedPulling="2026-01-29 12:52:35.901576941 +0000 UTC m=+2839.916707107" observedRunningTime="2026-01-29 12:52:36.372338615 +0000 UTC m=+2840.387468741" watchObservedRunningTime="2026-01-29 12:52:36.38028732 +0000 UTC m=+2840.395417446" Jan 29 12:52:41 crc kubenswrapper[4993]: I0129 12:52:41.906477 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:41 crc kubenswrapper[4993]: I0129 12:52:41.907379 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:41 crc kubenswrapper[4993]: I0129 12:52:41.971511 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:42 crc kubenswrapper[4993]: I0129 12:52:42.468484 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:42 crc kubenswrapper[4993]: I0129 12:52:42.529527 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.426060 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hvg8l" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="registry-server" containerID="cri-o://a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971" gracePeriod=2 Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.811694 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.897077 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4r9c\" (UniqueName: \"kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c\") pod \"bf38e4e1-adbd-4433-b65a-30a335742c4e\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.897152 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content\") pod \"bf38e4e1-adbd-4433-b65a-30a335742c4e\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.897578 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities\") pod \"bf38e4e1-adbd-4433-b65a-30a335742c4e\" (UID: \"bf38e4e1-adbd-4433-b65a-30a335742c4e\") " Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.899572 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities" (OuterVolumeSpecName: "utilities") pod "bf38e4e1-adbd-4433-b65a-30a335742c4e" (UID: "bf38e4e1-adbd-4433-b65a-30a335742c4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.907147 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c" (OuterVolumeSpecName: "kube-api-access-x4r9c") pod "bf38e4e1-adbd-4433-b65a-30a335742c4e" (UID: "bf38e4e1-adbd-4433-b65a-30a335742c4e"). InnerVolumeSpecName "kube-api-access-x4r9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.923607 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf38e4e1-adbd-4433-b65a-30a335742c4e" (UID: "bf38e4e1-adbd-4433-b65a-30a335742c4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.999487 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4r9c\" (UniqueName: \"kubernetes.io/projected/bf38e4e1-adbd-4433-b65a-30a335742c4e-kube-api-access-x4r9c\") on node \"crc\" DevicePath \"\"" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.999536 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:52:44 crc kubenswrapper[4993]: I0129 12:52:44.999558 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf38e4e1-adbd-4433-b65a-30a335742c4e-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.435615 4993 generic.go:334] "Generic (PLEG): container finished" podID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerID="a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971" exitCode=0 Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.435671 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerDied","Data":"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971"} Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.435694 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvg8l" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.435726 4993 scope.go:117] "RemoveContainer" containerID="a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.435709 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvg8l" event={"ID":"bf38e4e1-adbd-4433-b65a-30a335742c4e","Type":"ContainerDied","Data":"ec6c92a76f5ae463f3db10414bbbecd96f843e6eb8b43b2f027280d097f51b63"} Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.459397 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.461789 4993 scope.go:117] "RemoveContainer" containerID="eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.465498 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvg8l"] Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.475667 4993 scope.go:117] "RemoveContainer" containerID="f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.502404 4993 scope.go:117] "RemoveContainer" containerID="a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971" Jan 29 12:52:45 crc kubenswrapper[4993]: E0129 12:52:45.502807 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971\": container with ID starting with a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971 not found: ID does not exist" containerID="a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.502857 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971"} err="failed to get container status \"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971\": rpc error: code = NotFound desc = could not find container \"a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971\": container with ID starting with a5980bff9916db1d19e519a27d2ba80a8feb9505742c53438170a5158ab96971 not found: ID does not exist" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.502888 4993 scope.go:117] "RemoveContainer" containerID="eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90" Jan 29 12:52:45 crc kubenswrapper[4993]: E0129 12:52:45.503237 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90\": container with ID starting with eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90 not found: ID does not exist" containerID="eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.503281 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90"} err="failed to get container status \"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90\": rpc error: code = NotFound desc = could not find container \"eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90\": container with ID starting with eb802208bb534e6728e742ce486c58d2816443eca5122e704808e359d2c83a90 not found: ID does not exist" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.503309 4993 scope.go:117] "RemoveContainer" containerID="f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab" Jan 29 12:52:45 crc kubenswrapper[4993]: E0129 12:52:45.503607 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab\": container with ID starting with f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab not found: ID does not exist" containerID="f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab" Jan 29 12:52:45 crc kubenswrapper[4993]: I0129 12:52:45.503648 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab"} err="failed to get container status \"f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab\": rpc error: code = NotFound desc = could not find container \"f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab\": container with ID starting with f28a1fca40f2007a020ec63a35cd0d7409672e0dcfda935ae0596cbcee87cbab not found: ID does not exist" Jan 29 12:52:47 crc kubenswrapper[4993]: I0129 12:52:47.200219 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" path="/var/lib/kubelet/pods/bf38e4e1-adbd-4433-b65a-30a335742c4e/volumes" Jan 29 12:52:48 crc kubenswrapper[4993]: I0129 12:52:48.155540 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:52:48 crc kubenswrapper[4993]: I0129 12:52:48.155621 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.155518 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.156113 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.156167 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.156861 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.156925 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" gracePeriod=600 Jan 29 12:53:18 crc kubenswrapper[4993]: E0129 12:53:18.330508 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.708442 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" exitCode=0 Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.708497 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f"} Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.708542 4993 scope.go:117] "RemoveContainer" containerID="e4ed7005093830b385e57b6f94e1f728eb087a074b06a18967ecd09290a499be" Jan 29 12:53:18 crc kubenswrapper[4993]: I0129 12:53:18.710289 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:53:18 crc kubenswrapper[4993]: E0129 12:53:18.710643 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:53:32 crc kubenswrapper[4993]: I0129 12:53:32.189993 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:53:32 crc kubenswrapper[4993]: E0129 12:53:32.190689 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:53:47 crc kubenswrapper[4993]: I0129 12:53:47.195996 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:53:47 crc kubenswrapper[4993]: E0129 12:53:47.196599 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:53:59 crc kubenswrapper[4993]: I0129 12:53:59.190488 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:53:59 crc kubenswrapper[4993]: E0129 12:53:59.191278 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:54:10 crc kubenswrapper[4993]: I0129 12:54:10.191248 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:54:10 crc kubenswrapper[4993]: E0129 12:54:10.192540 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:54:21 crc kubenswrapper[4993]: I0129 12:54:21.191102 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:54:21 crc kubenswrapper[4993]: E0129 12:54:21.191738 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.821756 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:28 crc kubenswrapper[4993]: E0129 12:54:28.822676 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="registry-server" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.822692 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="registry-server" Jan 29 12:54:28 crc kubenswrapper[4993]: E0129 12:54:28.822709 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="extract-utilities" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.822718 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="extract-utilities" Jan 29 12:54:28 crc kubenswrapper[4993]: E0129 12:54:28.822740 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="extract-content" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.822750 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="extract-content" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.822932 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf38e4e1-adbd-4433-b65a-30a335742c4e" containerName="registry-server" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.824171 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.847017 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.909412 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.909722 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl8dm\" (UniqueName: \"kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:28 crc kubenswrapper[4993]: I0129 12:54:28.909857 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.011390 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.011464 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl8dm\" (UniqueName: \"kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.011510 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.012083 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.012132 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.030151 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl8dm\" (UniqueName: \"kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm\") pod \"community-operators-fsffz\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.152809 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:29 crc kubenswrapper[4993]: I0129 12:54:29.643480 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:30 crc kubenswrapper[4993]: I0129 12:54:30.257227 4993 generic.go:334] "Generic (PLEG): container finished" podID="f9048457-b96b-4142-ab14-b2e69cf96382" containerID="d615a8d2ad7b099bb3feb61966a7b821e08a5ea8b5f0dc36ac628514a8f8e728" exitCode=0 Jan 29 12:54:30 crc kubenswrapper[4993]: I0129 12:54:30.257346 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerDied","Data":"d615a8d2ad7b099bb3feb61966a7b821e08a5ea8b5f0dc36ac628514a8f8e728"} Jan 29 12:54:30 crc kubenswrapper[4993]: I0129 12:54:30.257554 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerStarted","Data":"fb1c6595bbca4d4a5fe7238616ce0ef80e126d42bf7ca043565df212a57fe454"} Jan 29 12:54:30 crc kubenswrapper[4993]: I0129 12:54:30.277069 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 12:54:32 crc kubenswrapper[4993]: I0129 12:54:32.191127 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:54:32 crc kubenswrapper[4993]: E0129 12:54:32.191806 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:54:32 crc kubenswrapper[4993]: I0129 12:54:32.282107 4993 generic.go:334] "Generic (PLEG): container finished" podID="f9048457-b96b-4142-ab14-b2e69cf96382" containerID="c05cf747f0aff3d8900daeb73d4c3b357bd877a503ce14356b746dcc51ec1595" exitCode=0 Jan 29 12:54:32 crc kubenswrapper[4993]: I0129 12:54:32.282226 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerDied","Data":"c05cf747f0aff3d8900daeb73d4c3b357bd877a503ce14356b746dcc51ec1595"} Jan 29 12:54:33 crc kubenswrapper[4993]: I0129 12:54:33.290539 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerStarted","Data":"3974cfde9d82931c162c6b6f6b55de703c998a2846a2e930cd5762e39becf961"} Jan 29 12:54:33 crc kubenswrapper[4993]: I0129 12:54:33.319631 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fsffz" podStartSLOduration=2.737412038 podStartE2EDuration="5.319613226s" podCreationTimestamp="2026-01-29 12:54:28 +0000 UTC" firstStartedPulling="2026-01-29 12:54:30.259262344 +0000 UTC m=+2954.274392510" lastFinishedPulling="2026-01-29 12:54:32.841463562 +0000 UTC m=+2956.856593698" observedRunningTime="2026-01-29 12:54:33.312438541 +0000 UTC m=+2957.327568667" watchObservedRunningTime="2026-01-29 12:54:33.319613226 +0000 UTC m=+2957.334743342" Jan 29 12:54:39 crc kubenswrapper[4993]: I0129 12:54:39.154425 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:39 crc kubenswrapper[4993]: I0129 12:54:39.155165 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:39 crc kubenswrapper[4993]: I0129 12:54:39.232949 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:39 crc kubenswrapper[4993]: I0129 12:54:39.373295 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:39 crc kubenswrapper[4993]: I0129 12:54:39.471271 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:41 crc kubenswrapper[4993]: I0129 12:54:41.354520 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fsffz" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="registry-server" containerID="cri-o://3974cfde9d82931c162c6b6f6b55de703c998a2846a2e930cd5762e39becf961" gracePeriod=2 Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.364677 4993 generic.go:334] "Generic (PLEG): container finished" podID="f9048457-b96b-4142-ab14-b2e69cf96382" containerID="3974cfde9d82931c162c6b6f6b55de703c998a2846a2e930cd5762e39becf961" exitCode=0 Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.364960 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerDied","Data":"3974cfde9d82931c162c6b6f6b55de703c998a2846a2e930cd5762e39becf961"} Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.517626 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.611754 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content\") pod \"f9048457-b96b-4142-ab14-b2e69cf96382\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.611936 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities\") pod \"f9048457-b96b-4142-ab14-b2e69cf96382\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.611979 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl8dm\" (UniqueName: \"kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm\") pod \"f9048457-b96b-4142-ab14-b2e69cf96382\" (UID: \"f9048457-b96b-4142-ab14-b2e69cf96382\") " Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.613399 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities" (OuterVolumeSpecName: "utilities") pod "f9048457-b96b-4142-ab14-b2e69cf96382" (UID: "f9048457-b96b-4142-ab14-b2e69cf96382"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.617205 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm" (OuterVolumeSpecName: "kube-api-access-xl8dm") pod "f9048457-b96b-4142-ab14-b2e69cf96382" (UID: "f9048457-b96b-4142-ab14-b2e69cf96382"). InnerVolumeSpecName "kube-api-access-xl8dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.713433 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:54:42 crc kubenswrapper[4993]: I0129 12:54:42.713461 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl8dm\" (UniqueName: \"kubernetes.io/projected/f9048457-b96b-4142-ab14-b2e69cf96382-kube-api-access-xl8dm\") on node \"crc\" DevicePath \"\"" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.190628 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:54:43 crc kubenswrapper[4993]: E0129 12:54:43.190855 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.220940 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9048457-b96b-4142-ab14-b2e69cf96382" (UID: "f9048457-b96b-4142-ab14-b2e69cf96382"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.322256 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9048457-b96b-4142-ab14-b2e69cf96382-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.374442 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsffz" event={"ID":"f9048457-b96b-4142-ab14-b2e69cf96382","Type":"ContainerDied","Data":"fb1c6595bbca4d4a5fe7238616ce0ef80e126d42bf7ca043565df212a57fe454"} Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.374491 4993 scope.go:117] "RemoveContainer" containerID="3974cfde9d82931c162c6b6f6b55de703c998a2846a2e930cd5762e39becf961" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.374584 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsffz" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.411337 4993 scope.go:117] "RemoveContainer" containerID="c05cf747f0aff3d8900daeb73d4c3b357bd877a503ce14356b746dcc51ec1595" Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.421659 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.430327 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fsffz"] Jan 29 12:54:43 crc kubenswrapper[4993]: I0129 12:54:43.448661 4993 scope.go:117] "RemoveContainer" containerID="d615a8d2ad7b099bb3feb61966a7b821e08a5ea8b5f0dc36ac628514a8f8e728" Jan 29 12:54:45 crc kubenswrapper[4993]: I0129 12:54:45.200074 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" path="/var/lib/kubelet/pods/f9048457-b96b-4142-ab14-b2e69cf96382/volumes" Jan 29 12:54:58 crc kubenswrapper[4993]: I0129 12:54:58.190875 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:54:58 crc kubenswrapper[4993]: E0129 12:54:58.191633 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:55:09 crc kubenswrapper[4993]: I0129 12:55:09.190945 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:55:09 crc kubenswrapper[4993]: E0129 12:55:09.192301 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:55:24 crc kubenswrapper[4993]: I0129 12:55:24.190910 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:55:24 crc kubenswrapper[4993]: E0129 12:55:24.191958 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:55:36 crc kubenswrapper[4993]: I0129 12:55:36.191037 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:55:36 crc kubenswrapper[4993]: E0129 12:55:36.191975 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:55:51 crc kubenswrapper[4993]: I0129 12:55:51.190268 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:55:51 crc kubenswrapper[4993]: E0129 12:55:51.191060 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:56:05 crc kubenswrapper[4993]: I0129 12:56:05.190047 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:56:05 crc kubenswrapper[4993]: E0129 12:56:05.190792 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.158507 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-97nqj/must-gather-69th9"] Jan 29 12:56:06 crc kubenswrapper[4993]: E0129 12:56:06.159201 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="extract-utilities" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.159226 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="extract-utilities" Jan 29 12:56:06 crc kubenswrapper[4993]: E0129 12:56:06.159239 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="extract-content" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.159250 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="extract-content" Jan 29 12:56:06 crc kubenswrapper[4993]: E0129 12:56:06.159260 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="registry-server" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.159268 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="registry-server" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.159460 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9048457-b96b-4142-ab14-b2e69cf96382" containerName="registry-server" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.160203 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.162133 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-97nqj"/"default-dockercfg-6gjhr" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.162884 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-97nqj"/"openshift-service-ca.crt" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.163227 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-97nqj"/"kube-root-ca.crt" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.182195 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-97nqj/must-gather-69th9"] Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.245114 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.245170 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kpmx\" (UniqueName: \"kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.346809 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.346887 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kpmx\" (UniqueName: \"kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.348154 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.365544 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kpmx\" (UniqueName: \"kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx\") pod \"must-gather-69th9\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.479257 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:56:06 crc kubenswrapper[4993]: I0129 12:56:06.922098 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-97nqj/must-gather-69th9"] Jan 29 12:56:06 crc kubenswrapper[4993]: W0129 12:56:06.932002 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e85de16_df5f_4527_a4e1_995ef92fc423.slice/crio-b862cb523a02bd00763e7f5adb0eb0fd78b0b09e183b813119af9d56f3425bea WatchSource:0}: Error finding container b862cb523a02bd00763e7f5adb0eb0fd78b0b09e183b813119af9d56f3425bea: Status 404 returned error can't find the container with id b862cb523a02bd00763e7f5adb0eb0fd78b0b09e183b813119af9d56f3425bea Jan 29 12:56:07 crc kubenswrapper[4993]: I0129 12:56:07.101467 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-97nqj/must-gather-69th9" event={"ID":"4e85de16-df5f-4527-a4e1-995ef92fc423","Type":"ContainerStarted","Data":"b862cb523a02bd00763e7f5adb0eb0fd78b0b09e183b813119af9d56f3425bea"} Jan 29 12:56:13 crc kubenswrapper[4993]: I0129 12:56:13.138982 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-97nqj/must-gather-69th9" event={"ID":"4e85de16-df5f-4527-a4e1-995ef92fc423","Type":"ContainerStarted","Data":"8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859"} Jan 29 12:56:13 crc kubenswrapper[4993]: I0129 12:56:13.139278 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-97nqj/must-gather-69th9" event={"ID":"4e85de16-df5f-4527-a4e1-995ef92fc423","Type":"ContainerStarted","Data":"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d"} Jan 29 12:56:13 crc kubenswrapper[4993]: I0129 12:56:13.200696 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-97nqj/must-gather-69th9" podStartSLOduration=1.773090877 podStartE2EDuration="7.200684283s" podCreationTimestamp="2026-01-29 12:56:06 +0000 UTC" firstStartedPulling="2026-01-29 12:56:06.935013711 +0000 UTC m=+3050.950143837" lastFinishedPulling="2026-01-29 12:56:12.362607117 +0000 UTC m=+3056.377737243" observedRunningTime="2026-01-29 12:56:13.197173955 +0000 UTC m=+3057.212304081" watchObservedRunningTime="2026-01-29 12:56:13.200684283 +0000 UTC m=+3057.215814409" Jan 29 12:56:16 crc kubenswrapper[4993]: I0129 12:56:16.190704 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:56:16 crc kubenswrapper[4993]: E0129 12:56:16.191266 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:56:28 crc kubenswrapper[4993]: I0129 12:56:28.191115 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:56:28 crc kubenswrapper[4993]: E0129 12:56:28.191982 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:56:39 crc kubenswrapper[4993]: I0129 12:56:39.190632 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:56:39 crc kubenswrapper[4993]: E0129 12:56:39.191269 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:56:50 crc kubenswrapper[4993]: I0129 12:56:50.190415 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:56:50 crc kubenswrapper[4993]: E0129 12:56:50.191213 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:57:04 crc kubenswrapper[4993]: I0129 12:57:04.190288 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:57:04 crc kubenswrapper[4993]: E0129 12:57:04.190866 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.226593 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-657667746d-hp6j7_523bfe53-fa56-448a-9bfe-bc268342e493/manager/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.360006 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/util/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.523241 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/pull/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.552396 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/pull/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.593442 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/util/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.719179 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/pull/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.756372 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/util/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.774116 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c4e852f88ee0550736c43fc6f039704d50126811cd6410eb28002ff66fhpx4k_9fdf2bc7-fef0-4ada-b491-f963835bb1cb/extract/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.897557 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7595cf584-5zt8x_02adb9a3-e33a-4ad6-8563-078804260425/manager/0.log" Jan 29 12:57:11 crc kubenswrapper[4993]: I0129 12:57:11.962461 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-55d5d5f8ff-846wm_85859ee8-efc0-40b9-a3a0-9ad36c72745f/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.091238 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6db5dbd896-qdnx9_c329c584-dda3-408c-9646-1f7ad46fcff2/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.170614 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5499bccc75-9gmx8_3670c965-03c3-40d0-9990-daacad669caa/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.312323 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-cp6x6_4ed2c1a4-5753-4802-bbc4-53d89977d7a9/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.378138 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-c9n87_66e97264-1497-4d2d-968f-e2edc9dc9017/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.485254 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-56cb7c4b4c-925f9_dcb6de55-31e4-44fc-8e36-d3046fe4d8b2/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.639104 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6475bdcbc4-8m42n_552f4e86-01b8-4c1c-bfdc-e6e7ad197682/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.665564 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-77bb7ffb8c-xlnqs_c382415b-73b1-4fe9-a4f0-f828627afb02/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.775575 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-48fq8_39304887-7ccc-4b57-95d6-64f886c15e54/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.866744 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-55df775b69-2jnz5_4b6f17b9-04e0-44b0-bd5d-7cb91cc9efed/manager/0.log" Jan 29 12:57:12 crc kubenswrapper[4993]: I0129 12:57:12.970101 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5ccd5b7f8f-c6gdt_aa36f54c-d314-4086-a64e-acec3c80c8bd/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.023848 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6b855b4fc4-jrwl8_615c8615-803e-4997-819a-9eecd22cdbee/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.173109 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-59c4b45c4d8t8gs_30c2b0bb-dbda-49d5-92f6-567703290c05/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.402058 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65dc8f5954-7dnzd_6c7a911f-47c8-4fab-adb0-7d179c090b29/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.404809 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-59c8666fb5-c6hrf_b1d125da-2f26-485e-9010-fd330ee694cd/operator/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.736947 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-7m6nn_466ba75a-ab3e-481b-8554-fcf15f73bea6/registry-server/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.741483 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-sqf57_10455fa4-3adf-4e84-8a37-ae1241d62df6/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.932350 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-2xcgn_66cb8608-8dd0-4304-9245-8c4468d7c6eb/manager/0.log" Jan 29 12:57:13 crc kubenswrapper[4993]: I0129 12:57:13.945839 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-pxqq5_181f3d8c-9fd7-4df6-806a-8d04ac0ff597/operator/0.log" Jan 29 12:57:14 crc kubenswrapper[4993]: I0129 12:57:14.126459 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-c95fd9dc5-fqltq_c3df3d31-0166-4aaa-b6a8-98865bc4d99c/manager/0.log" Jan 29 12:57:14 crc kubenswrapper[4993]: I0129 12:57:14.129569 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6f7455757b-ls4r8_0a0fde08-a591-4fb8-a872-c0a4cd7e2758/manager/0.log" Jan 29 12:57:14 crc kubenswrapper[4993]: I0129 12:57:14.323413 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-m9xms_be15d7ad-6e9d-40fe-845e-626f4b1dc766/manager/0.log" Jan 29 12:57:14 crc kubenswrapper[4993]: I0129 12:57:14.360688 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-56b5dc77fd-v57np_da5d8a04-0e63-4b7a-b1b7-460a97473fce/manager/0.log" Jan 29 12:57:19 crc kubenswrapper[4993]: I0129 12:57:19.190398 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:57:19 crc kubenswrapper[4993]: E0129 12:57:19.191109 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:57:32 crc kubenswrapper[4993]: I0129 12:57:32.190306 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:57:32 crc kubenswrapper[4993]: E0129 12:57:32.192778 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:57:32 crc kubenswrapper[4993]: I0129 12:57:32.605861 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-c4z7l_2becf916-8ac8-48ad-ba1c-e679036dd069/control-plane-machine-set-operator/0.log" Jan 29 12:57:32 crc kubenswrapper[4993]: I0129 12:57:32.769253 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wfg4q_3ef5ad33-4ec5-4674-a517-8ade25ce2569/kube-rbac-proxy/0.log" Jan 29 12:57:32 crc kubenswrapper[4993]: I0129 12:57:32.774795 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wfg4q_3ef5ad33-4ec5-4674-a517-8ade25ce2569/machine-api-operator/0.log" Jan 29 12:57:44 crc kubenswrapper[4993]: I0129 12:57:44.593399 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-nr47b_bf061abb-955a-43c6-860a-cb439391540e/cert-manager-controller/0.log" Jan 29 12:57:44 crc kubenswrapper[4993]: I0129 12:57:44.721114 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-7wxc2_bfa834cf-9989-403e-b3f8-3ec8b7b62204/cert-manager-cainjector/0.log" Jan 29 12:57:44 crc kubenswrapper[4993]: I0129 12:57:44.788985 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-z9fkp_58863fed-2920-49fd-a1bc-6a5f8a771783/cert-manager-webhook/0.log" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.092700 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.094404 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.100079 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.100325 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.100401 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nmd7\" (UniqueName: \"kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.111107 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.201416 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.201514 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.201548 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nmd7\" (UniqueName: \"kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.201999 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.202059 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.223204 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nmd7\" (UniqueName: \"kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7\") pod \"certified-operators-mmtqb\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.415632 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:45 crc kubenswrapper[4993]: I0129 12:57:45.996005 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:46 crc kubenswrapper[4993]: I0129 12:57:46.985757 4993 generic.go:334] "Generic (PLEG): container finished" podID="332525f7-229a-4116-8d7d-219c727bb9c9" containerID="df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87" exitCode=0 Jan 29 12:57:46 crc kubenswrapper[4993]: I0129 12:57:46.987560 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerDied","Data":"df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87"} Jan 29 12:57:46 crc kubenswrapper[4993]: I0129 12:57:46.987595 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerStarted","Data":"babc62247b4ec75d012275690b2eef6c4d42633a3b17774f0d615db6d669bdfc"} Jan 29 12:57:47 crc kubenswrapper[4993]: I0129 12:57:47.194869 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:57:47 crc kubenswrapper[4993]: E0129 12:57:47.195042 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:57:48 crc kubenswrapper[4993]: I0129 12:57:48.999076 4993 generic.go:334] "Generic (PLEG): container finished" podID="332525f7-229a-4116-8d7d-219c727bb9c9" containerID="03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422" exitCode=0 Jan 29 12:57:48 crc kubenswrapper[4993]: I0129 12:57:48.999135 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerDied","Data":"03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422"} Jan 29 12:57:50 crc kubenswrapper[4993]: I0129 12:57:50.013082 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerStarted","Data":"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007"} Jan 29 12:57:50 crc kubenswrapper[4993]: I0129 12:57:50.038914 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mmtqb" podStartSLOduration=2.226180007 podStartE2EDuration="5.038894652s" podCreationTimestamp="2026-01-29 12:57:45 +0000 UTC" firstStartedPulling="2026-01-29 12:57:46.989614685 +0000 UTC m=+3151.004744811" lastFinishedPulling="2026-01-29 12:57:49.80232933 +0000 UTC m=+3153.817459456" observedRunningTime="2026-01-29 12:57:50.034350957 +0000 UTC m=+3154.049481123" watchObservedRunningTime="2026-01-29 12:57:50.038894652 +0000 UTC m=+3154.054024778" Jan 29 12:57:55 crc kubenswrapper[4993]: I0129 12:57:55.416538 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:55 crc kubenswrapper[4993]: I0129 12:57:55.416995 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:55 crc kubenswrapper[4993]: I0129 12:57:55.466758 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:56 crc kubenswrapper[4993]: I0129 12:57:56.098643 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:56 crc kubenswrapper[4993]: I0129 12:57:56.155002 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.284988 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-vn2fg_24b06982-4e48-47f3-84ed-8fd0f34b1f43/nmstate-console-plugin/0.log" Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.431176 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-tzsvj_dd99b0bb-57cb-4034-bf70-c27b85d936d2/nmstate-handler/0.log" Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.474883 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvw7j_b407b3c8-24b5-400b-967c-f8145f41bc4b/kube-rbac-proxy/0.log" Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.553435 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvw7j_b407b3c8-24b5-400b-967c-f8145f41bc4b/nmstate-metrics/0.log" Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.657873 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-s54j9_5719c06f-4107-47e6-af46-cc7fcc25d3d8/nmstate-operator/0.log" Jan 29 12:57:57 crc kubenswrapper[4993]: I0129 12:57:57.740431 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-2hphw_588bc834-7403-44b7-8152-7a18d3da5e92/nmstate-webhook/0.log" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.066403 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mmtqb" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="registry-server" containerID="cri-o://0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007" gracePeriod=2 Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.468272 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.594435 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content\") pod \"332525f7-229a-4116-8d7d-219c727bb9c9\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.594475 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nmd7\" (UniqueName: \"kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7\") pod \"332525f7-229a-4116-8d7d-219c727bb9c9\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.594498 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities\") pod \"332525f7-229a-4116-8d7d-219c727bb9c9\" (UID: \"332525f7-229a-4116-8d7d-219c727bb9c9\") " Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.595498 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities" (OuterVolumeSpecName: "utilities") pod "332525f7-229a-4116-8d7d-219c727bb9c9" (UID: "332525f7-229a-4116-8d7d-219c727bb9c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.600403 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7" (OuterVolumeSpecName: "kube-api-access-7nmd7") pod "332525f7-229a-4116-8d7d-219c727bb9c9" (UID: "332525f7-229a-4116-8d7d-219c727bb9c9"). InnerVolumeSpecName "kube-api-access-7nmd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.645154 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "332525f7-229a-4116-8d7d-219c727bb9c9" (UID: "332525f7-229a-4116-8d7d-219c727bb9c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.696109 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.696148 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nmd7\" (UniqueName: \"kubernetes.io/projected/332525f7-229a-4116-8d7d-219c727bb9c9-kube-api-access-7nmd7\") on node \"crc\" DevicePath \"\"" Jan 29 12:57:58 crc kubenswrapper[4993]: I0129 12:57:58.696163 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332525f7-229a-4116-8d7d-219c727bb9c9-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.077429 4993 generic.go:334] "Generic (PLEG): container finished" podID="332525f7-229a-4116-8d7d-219c727bb9c9" containerID="0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007" exitCode=0 Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.077478 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtqb" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.077487 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerDied","Data":"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007"} Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.077526 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtqb" event={"ID":"332525f7-229a-4116-8d7d-219c727bb9c9","Type":"ContainerDied","Data":"babc62247b4ec75d012275690b2eef6c4d42633a3b17774f0d615db6d669bdfc"} Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.077555 4993 scope.go:117] "RemoveContainer" containerID="0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.099687 4993 scope.go:117] "RemoveContainer" containerID="03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.125147 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.131877 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mmtqb"] Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.133552 4993 scope.go:117] "RemoveContainer" containerID="df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.152819 4993 scope.go:117] "RemoveContainer" containerID="0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007" Jan 29 12:57:59 crc kubenswrapper[4993]: E0129 12:57:59.153290 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007\": container with ID starting with 0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007 not found: ID does not exist" containerID="0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.153327 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007"} err="failed to get container status \"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007\": rpc error: code = NotFound desc = could not find container \"0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007\": container with ID starting with 0b693c3c004b1840db43e13c07d26844a1351391a564005be5041c0dde7f8007 not found: ID does not exist" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.153348 4993 scope.go:117] "RemoveContainer" containerID="03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422" Jan 29 12:57:59 crc kubenswrapper[4993]: E0129 12:57:59.153724 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422\": container with ID starting with 03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422 not found: ID does not exist" containerID="03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.153758 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422"} err="failed to get container status \"03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422\": rpc error: code = NotFound desc = could not find container \"03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422\": container with ID starting with 03dd67c95844fdc2aa0b0e67d72332c02443c82b293fb3e55719e013f4de3422 not found: ID does not exist" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.153785 4993 scope.go:117] "RemoveContainer" containerID="df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87" Jan 29 12:57:59 crc kubenswrapper[4993]: E0129 12:57:59.154078 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87\": container with ID starting with df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87 not found: ID does not exist" containerID="df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.154102 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87"} err="failed to get container status \"df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87\": rpc error: code = NotFound desc = could not find container \"df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87\": container with ID starting with df1da185de6f337845abe75defc63a4576b16fa20ce38bef3dade0df43869f87 not found: ID does not exist" Jan 29 12:57:59 crc kubenswrapper[4993]: I0129 12:57:59.208469 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" path="/var/lib/kubelet/pods/332525f7-229a-4116-8d7d-219c727bb9c9/volumes" Jan 29 12:58:00 crc kubenswrapper[4993]: I0129 12:58:00.191105 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:58:00 crc kubenswrapper[4993]: E0129 12:58:00.191989 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:58:15 crc kubenswrapper[4993]: I0129 12:58:15.191309 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:58:15 crc kubenswrapper[4993]: E0129 12:58:15.193933 4993 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bskq4_openshift-machine-config-operator(b1df6750-623d-4dda-83cd-dec067f8fc1a)\"" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.438426 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-zlzsh_36e0eaff-f539-449a-8cab-3a8e67a503da/kube-rbac-proxy/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.566441 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-zlzsh_36e0eaff-f539-449a-8cab-3a8e67a503da/controller/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.665459 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-frr-files/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.828358 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-metrics/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.831406 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-frr-files/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.842911 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-reloader/0.log" Jan 29 12:58:22 crc kubenswrapper[4993]: I0129 12:58:22.843254 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-reloader/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.016402 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-frr-files/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.017982 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-metrics/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.054460 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-reloader/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.068697 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-metrics/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.237932 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-reloader/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.246297 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-metrics/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.247421 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/controller/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.252004 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/cp-frr-files/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.427825 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/kube-rbac-proxy/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.432446 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/kube-rbac-proxy-frr/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.465214 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/frr-metrics/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.636993 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-m5726_8f6aec89-4069-4b04-8258-d9fbeb62bd79/frr-k8s-webhook-server/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.654493 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/reloader/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.656748 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sjhwc_ec787c89-bd58-470e-942e-b9775f1ee08e/frr/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.852106 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5496d86f-5p52h_a7d00ccb-f451-43bb-8280-08604b5ddf86/manager/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.880090 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75647f5f8d-xh6r2_89b47e24-436a-4914-b4ff-c9767a62ac58/webhook-server/0.log" Jan 29 12:58:23 crc kubenswrapper[4993]: I0129 12:58:23.992118 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-99rlj_5c25e8c1-f850-435e-956c-92cb3e13be64/kube-rbac-proxy/0.log" Jan 29 12:58:24 crc kubenswrapper[4993]: I0129 12:58:24.157697 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-99rlj_5c25e8c1-f850-435e-956c-92cb3e13be64/speaker/0.log" Jan 29 12:58:29 crc kubenswrapper[4993]: I0129 12:58:29.190273 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 12:58:30 crc kubenswrapper[4993]: I0129 12:58:30.313236 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"07bbb89c41b9031371d9296fcb69777efc4d723b1c8c6f8a1707b9368625a7eb"} Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.452908 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/util/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.599274 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/util/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.606804 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/pull/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.624519 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/pull/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.791342 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/util/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.807310 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/pull/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.810974 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcqrqkq_c8952af7-d5e0-4a95-b584-347ba5d87a16/extract/0.log" Jan 29 12:58:36 crc kubenswrapper[4993]: I0129 12:58:36.982426 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/util/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.283096 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/pull/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.294888 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/util/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.307597 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/pull/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.440554 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/util/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.440919 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/pull/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.450585 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713mvp26_580e9990-2a5c-4761-b9a6-1111d922725c/extract/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.623882 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/util/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.840326 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/pull/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.854491 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/util/0.log" Jan 29 12:58:37 crc kubenswrapper[4993]: I0129 12:58:37.873292 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/pull/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.061419 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/util/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.066002 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/extract/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.127282 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5fnkwl_cbc114ee-d9f7-4bdb-a104-6e8b1c7aaf6a/pull/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.261975 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-utilities/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.445169 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-content/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.452968 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-content/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.453099 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-utilities/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.654609 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-utilities/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.660288 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/extract-content/0.log" Jan 29 12:58:38 crc kubenswrapper[4993]: I0129 12:58:38.878411 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-utilities/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.037785 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jbcpr_a175a8e3-3022-4953-9419-4fc74c9fb0e0/registry-server/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.039258 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.042310 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-utilities/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.090519 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.277988 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.320646 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/extract-utilities/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.477946 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qvrlt_2f0e3bcf-c987-4ef4-9055-31b947d2807c/marketplace-operator/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.587531 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-utilities/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.651550 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w5fq_98b3b4b4-59d6-4e70-ab5b-f763edbc691a/registry-server/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.763249 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.776616 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-utilities/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.801378 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.978496 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-content/0.log" Jan 29 12:58:39 crc kubenswrapper[4993]: I0129 12:58:39.983774 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/extract-utilities/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.068164 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sjktr_768166e9-0d56-4f39-9901-0da8ea9bd051/registry-server/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.196879 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-utilities/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.358444 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-utilities/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.393194 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-content/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.428555 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-content/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.524393 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-content/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.533354 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/extract-utilities/0.log" Jan 29 12:58:40 crc kubenswrapper[4993]: I0129 12:58:40.962858 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fpf2p_f36b3ae1-b711-44a3-900e-2c51754a7109/registry-server/0.log" Jan 29 12:59:48 crc kubenswrapper[4993]: I0129 12:59:48.916824 4993 generic.go:334] "Generic (PLEG): container finished" podID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerID="53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d" exitCode=0 Jan 29 12:59:48 crc kubenswrapper[4993]: I0129 12:59:48.916906 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-97nqj/must-gather-69th9" event={"ID":"4e85de16-df5f-4527-a4e1-995ef92fc423","Type":"ContainerDied","Data":"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d"} Jan 29 12:59:48 crc kubenswrapper[4993]: I0129 12:59:48.918101 4993 scope.go:117] "RemoveContainer" containerID="53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d" Jan 29 12:59:49 crc kubenswrapper[4993]: I0129 12:59:49.829230 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-97nqj_must-gather-69th9_4e85de16-df5f-4527-a4e1-995ef92fc423/gather/0.log" Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.577018 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-97nqj/must-gather-69th9"] Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.577794 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-97nqj/must-gather-69th9" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="copy" containerID="cri-o://8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859" gracePeriod=2 Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.583304 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-97nqj/must-gather-69th9"] Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.948003 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-97nqj_must-gather-69th9_4e85de16-df5f-4527-a4e1-995ef92fc423/copy/0.log" Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.948607 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.979331 4993 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-97nqj_must-gather-69th9_4e85de16-df5f-4527-a4e1-995ef92fc423/copy/0.log" Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.980101 4993 generic.go:334] "Generic (PLEG): container finished" podID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerID="8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859" exitCode=143 Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.980167 4993 scope.go:117] "RemoveContainer" containerID="8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859" Jan 29 12:59:56 crc kubenswrapper[4993]: I0129 12:59:56.980178 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-97nqj/must-gather-69th9" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.001237 4993 scope.go:117] "RemoveContainer" containerID="53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.049753 4993 scope.go:117] "RemoveContainer" containerID="8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.050026 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kpmx\" (UniqueName: \"kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx\") pod \"4e85de16-df5f-4527-a4e1-995ef92fc423\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.050077 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output\") pod \"4e85de16-df5f-4527-a4e1-995ef92fc423\" (UID: \"4e85de16-df5f-4527-a4e1-995ef92fc423\") " Jan 29 12:59:57 crc kubenswrapper[4993]: E0129 12:59:57.050105 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859\": container with ID starting with 8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859 not found: ID does not exist" containerID="8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.050146 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859"} err="failed to get container status \"8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859\": rpc error: code = NotFound desc = could not find container \"8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859\": container with ID starting with 8bf3b20d8a323c7cb2b6e8c313729a6395f2a73996627deaa965eeff4a1e4859 not found: ID does not exist" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.050172 4993 scope.go:117] "RemoveContainer" containerID="53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d" Jan 29 12:59:57 crc kubenswrapper[4993]: E0129 12:59:57.050642 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d\": container with ID starting with 53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d not found: ID does not exist" containerID="53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.050675 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d"} err="failed to get container status \"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d\": rpc error: code = NotFound desc = could not find container \"53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d\": container with ID starting with 53c64845a518282c4f8d28e8002b0913265b4ee6d0f27c22a0aa80de0f387c2d not found: ID does not exist" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.055342 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx" (OuterVolumeSpecName: "kube-api-access-6kpmx") pod "4e85de16-df5f-4527-a4e1-995ef92fc423" (UID: "4e85de16-df5f-4527-a4e1-995ef92fc423"). InnerVolumeSpecName "kube-api-access-6kpmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.128755 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4e85de16-df5f-4527-a4e1-995ef92fc423" (UID: "4e85de16-df5f-4527-a4e1-995ef92fc423"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.152111 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kpmx\" (UniqueName: \"kubernetes.io/projected/4e85de16-df5f-4527-a4e1-995ef92fc423-kube-api-access-6kpmx\") on node \"crc\" DevicePath \"\"" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.152160 4993 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4e85de16-df5f-4527-a4e1-995ef92fc423-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 29 12:59:57 crc kubenswrapper[4993]: I0129 12:59:57.201535 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" path="/var/lib/kubelet/pods/4e85de16-df5f-4527-a4e1-995ef92fc423/volumes" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.149614 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l"] Jan 29 13:00:00 crc kubenswrapper[4993]: E0129 13:00:00.150635 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="gather" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150661 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="gather" Jan 29 13:00:00 crc kubenswrapper[4993]: E0129 13:00:00.150684 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="extract-content" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150691 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="extract-content" Jan 29 13:00:00 crc kubenswrapper[4993]: E0129 13:00:00.150698 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="extract-utilities" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150707 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="extract-utilities" Jan 29 13:00:00 crc kubenswrapper[4993]: E0129 13:00:00.150718 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="copy" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150724 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="copy" Jan 29 13:00:00 crc kubenswrapper[4993]: E0129 13:00:00.150739 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="registry-server" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150745 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="registry-server" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150874 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="copy" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150893 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="332525f7-229a-4116-8d7d-219c727bb9c9" containerName="registry-server" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.150905 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e85de16-df5f-4527-a4e1-995ef92fc423" containerName="gather" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.151446 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.153333 4993 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.153643 4993 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.162540 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l"] Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.291862 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.292287 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.292606 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l79q6\" (UniqueName: \"kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.394173 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.394297 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l79q6\" (UniqueName: \"kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.394328 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.395169 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.399113 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.415400 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l79q6\" (UniqueName: \"kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6\") pod \"collect-profiles-29494860-gjl5l\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.482300 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:00 crc kubenswrapper[4993]: I0129 13:00:00.879750 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l"] Jan 29 13:00:01 crc kubenswrapper[4993]: I0129 13:00:01.024392 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" event={"ID":"50ced67f-594e-4596-8d59-f702b35d8692","Type":"ContainerStarted","Data":"57f49c62bd5db379f73d65ff92f8666fcf701d65aa2da245231c1c128d4d22e0"} Jan 29 13:00:02 crc kubenswrapper[4993]: I0129 13:00:02.033516 4993 generic.go:334] "Generic (PLEG): container finished" podID="50ced67f-594e-4596-8d59-f702b35d8692" containerID="26b82cae9b097ec8b88533e13cfc5fa2ea714f6cd9472a264d3bca7951510ba0" exitCode=0 Jan 29 13:00:02 crc kubenswrapper[4993]: I0129 13:00:02.033585 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" event={"ID":"50ced67f-594e-4596-8d59-f702b35d8692","Type":"ContainerDied","Data":"26b82cae9b097ec8b88533e13cfc5fa2ea714f6cd9472a264d3bca7951510ba0"} Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.302267 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.448334 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume\") pod \"50ced67f-594e-4596-8d59-f702b35d8692\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.448732 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume\") pod \"50ced67f-594e-4596-8d59-f702b35d8692\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.448770 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l79q6\" (UniqueName: \"kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6\") pod \"50ced67f-594e-4596-8d59-f702b35d8692\" (UID: \"50ced67f-594e-4596-8d59-f702b35d8692\") " Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.449273 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume" (OuterVolumeSpecName: "config-volume") pod "50ced67f-594e-4596-8d59-f702b35d8692" (UID: "50ced67f-594e-4596-8d59-f702b35d8692"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.461414 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "50ced67f-594e-4596-8d59-f702b35d8692" (UID: "50ced67f-594e-4596-8d59-f702b35d8692"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.462503 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6" (OuterVolumeSpecName: "kube-api-access-l79q6") pod "50ced67f-594e-4596-8d59-f702b35d8692" (UID: "50ced67f-594e-4596-8d59-f702b35d8692"). InnerVolumeSpecName "kube-api-access-l79q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.550998 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l79q6\" (UniqueName: \"kubernetes.io/projected/50ced67f-594e-4596-8d59-f702b35d8692-kube-api-access-l79q6\") on node \"crc\" DevicePath \"\"" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.551036 4993 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50ced67f-594e-4596-8d59-f702b35d8692-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 13:00:03 crc kubenswrapper[4993]: I0129 13:00:03.551048 4993 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50ced67f-594e-4596-8d59-f702b35d8692-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 13:00:04 crc kubenswrapper[4993]: I0129 13:00:04.050316 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" event={"ID":"50ced67f-594e-4596-8d59-f702b35d8692","Type":"ContainerDied","Data":"57f49c62bd5db379f73d65ff92f8666fcf701d65aa2da245231c1c128d4d22e0"} Jan 29 13:00:04 crc kubenswrapper[4993]: I0129 13:00:04.050355 4993 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57f49c62bd5db379f73d65ff92f8666fcf701d65aa2da245231c1c128d4d22e0" Jan 29 13:00:04 crc kubenswrapper[4993]: I0129 13:00:04.050383 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494860-gjl5l" Jan 29 13:00:04 crc kubenswrapper[4993]: I0129 13:00:04.367519 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49"] Jan 29 13:00:04 crc kubenswrapper[4993]: I0129 13:00:04.373247 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494815-2xw49"] Jan 29 13:00:05 crc kubenswrapper[4993]: I0129 13:00:05.198515 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d48e225a-e472-497b-b1af-08f427868ed8" path="/var/lib/kubelet/pods/d48e225a-e472-497b-b1af-08f427868ed8/volumes" Jan 29 13:00:18 crc kubenswrapper[4993]: I0129 13:00:18.528079 4993 scope.go:117] "RemoveContainer" containerID="d257f61af223ff42f9577780254ceb22e7311ca2c78492fdefdc2540d941ad71" Jan 29 13:00:48 crc kubenswrapper[4993]: I0129 13:00:48.156023 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 13:00:48 crc kubenswrapper[4993]: I0129 13:00:48.156531 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 13:01:18 crc kubenswrapper[4993]: I0129 13:01:18.155634 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 13:01:18 crc kubenswrapper[4993]: I0129 13:01:18.156249 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.156172 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.156786 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.156844 4993 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.157764 4993 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07bbb89c41b9031371d9296fcb69777efc4d723b1c8c6f8a1707b9368625a7eb"} pod="openshift-machine-config-operator/machine-config-daemon-bskq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.158003 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" containerID="cri-o://07bbb89c41b9031371d9296fcb69777efc4d723b1c8c6f8a1707b9368625a7eb" gracePeriod=600 Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.856147 4993 generic.go:334] "Generic (PLEG): container finished" podID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerID="07bbb89c41b9031371d9296fcb69777efc4d723b1c8c6f8a1707b9368625a7eb" exitCode=0 Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.856207 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerDied","Data":"07bbb89c41b9031371d9296fcb69777efc4d723b1c8c6f8a1707b9368625a7eb"} Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.856840 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" event={"ID":"b1df6750-623d-4dda-83cd-dec067f8fc1a","Type":"ContainerStarted","Data":"df5896b9fbdd6e95248888f05c3ee4c6f7c4bbc2f8fc7f310a5be2e437ab1e6f"} Jan 29 13:01:48 crc kubenswrapper[4993]: I0129 13:01:48.856893 4993 scope.go:117] "RemoveContainer" containerID="4b083b38254823934850e26f94c53212ee99d354a37cfd52885f16c6ffef9d6f" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.304526 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:01:55 crc kubenswrapper[4993]: E0129 13:01:55.305424 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50ced67f-594e-4596-8d59-f702b35d8692" containerName="collect-profiles" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.305441 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="50ced67f-594e-4596-8d59-f702b35d8692" containerName="collect-profiles" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.305647 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="50ced67f-594e-4596-8d59-f702b35d8692" containerName="collect-profiles" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.306679 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.335698 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.463286 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms4p2\" (UniqueName: \"kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.463450 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.463655 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.565064 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.565173 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms4p2\" (UniqueName: \"kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.565272 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.565846 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.566136 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.585748 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms4p2\" (UniqueName: \"kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2\") pod \"redhat-operators-t82rw\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:55 crc kubenswrapper[4993]: I0129 13:01:55.651578 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:01:56 crc kubenswrapper[4993]: I0129 13:01:56.076608 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:01:56 crc kubenswrapper[4993]: W0129 13:01:56.080156 4993 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9176028a_a581_4e83_823b_a519d2907240.slice/crio-ddf7ed8dc4b01d9f684163b45baf99e8633ff3a88563c8fdf0b539beb23defaf WatchSource:0}: Error finding container ddf7ed8dc4b01d9f684163b45baf99e8633ff3a88563c8fdf0b539beb23defaf: Status 404 returned error can't find the container with id ddf7ed8dc4b01d9f684163b45baf99e8633ff3a88563c8fdf0b539beb23defaf Jan 29 13:01:56 crc kubenswrapper[4993]: I0129 13:01:56.921706 4993 generic.go:334] "Generic (PLEG): container finished" podID="9176028a-a581-4e83-823b-a519d2907240" containerID="f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b" exitCode=0 Jan 29 13:01:56 crc kubenswrapper[4993]: I0129 13:01:56.921751 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerDied","Data":"f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b"} Jan 29 13:01:56 crc kubenswrapper[4993]: I0129 13:01:56.921972 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerStarted","Data":"ddf7ed8dc4b01d9f684163b45baf99e8633ff3a88563c8fdf0b539beb23defaf"} Jan 29 13:01:56 crc kubenswrapper[4993]: I0129 13:01:56.923332 4993 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 13:01:57 crc kubenswrapper[4993]: I0129 13:01:57.929818 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerStarted","Data":"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29"} Jan 29 13:01:58 crc kubenswrapper[4993]: I0129 13:01:58.939308 4993 generic.go:334] "Generic (PLEG): container finished" podID="9176028a-a581-4e83-823b-a519d2907240" containerID="7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29" exitCode=0 Jan 29 13:01:58 crc kubenswrapper[4993]: I0129 13:01:58.939359 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerDied","Data":"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29"} Jan 29 13:01:59 crc kubenswrapper[4993]: I0129 13:01:59.947105 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerStarted","Data":"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc"} Jan 29 13:01:59 crc kubenswrapper[4993]: I0129 13:01:59.966257 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t82rw" podStartSLOduration=2.171032995 podStartE2EDuration="4.966241197s" podCreationTimestamp="2026-01-29 13:01:55 +0000 UTC" firstStartedPulling="2026-01-29 13:01:56.923072471 +0000 UTC m=+3400.938202597" lastFinishedPulling="2026-01-29 13:01:59.718280673 +0000 UTC m=+3403.733410799" observedRunningTime="2026-01-29 13:01:59.965516326 +0000 UTC m=+3403.980646442" watchObservedRunningTime="2026-01-29 13:01:59.966241197 +0000 UTC m=+3403.981371323" Jan 29 13:02:05 crc kubenswrapper[4993]: I0129 13:02:05.652268 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:05 crc kubenswrapper[4993]: I0129 13:02:05.654600 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:05 crc kubenswrapper[4993]: I0129 13:02:05.698197 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:06 crc kubenswrapper[4993]: I0129 13:02:06.040946 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:06 crc kubenswrapper[4993]: I0129 13:02:06.087395 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.001331 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t82rw" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="registry-server" containerID="cri-o://460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc" gracePeriod=2 Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.408304 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.544924 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms4p2\" (UniqueName: \"kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2\") pod \"9176028a-a581-4e83-823b-a519d2907240\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.545242 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content\") pod \"9176028a-a581-4e83-823b-a519d2907240\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.545331 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities\") pod \"9176028a-a581-4e83-823b-a519d2907240\" (UID: \"9176028a-a581-4e83-823b-a519d2907240\") " Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.546376 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities" (OuterVolumeSpecName: "utilities") pod "9176028a-a581-4e83-823b-a519d2907240" (UID: "9176028a-a581-4e83-823b-a519d2907240"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.553028 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2" (OuterVolumeSpecName: "kube-api-access-ms4p2") pod "9176028a-a581-4e83-823b-a519d2907240" (UID: "9176028a-a581-4e83-823b-a519d2907240"). InnerVolumeSpecName "kube-api-access-ms4p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.649225 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 13:02:08 crc kubenswrapper[4993]: I0129 13:02:08.649263 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms4p2\" (UniqueName: \"kubernetes.io/projected/9176028a-a581-4e83-823b-a519d2907240-kube-api-access-ms4p2\") on node \"crc\" DevicePath \"\"" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.014957 4993 generic.go:334] "Generic (PLEG): container finished" podID="9176028a-a581-4e83-823b-a519d2907240" containerID="460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc" exitCode=0 Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.015023 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerDied","Data":"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc"} Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.015376 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t82rw" event={"ID":"9176028a-a581-4e83-823b-a519d2907240","Type":"ContainerDied","Data":"ddf7ed8dc4b01d9f684163b45baf99e8633ff3a88563c8fdf0b539beb23defaf"} Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.015411 4993 scope.go:117] "RemoveContainer" containerID="460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.015028 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t82rw" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.036889 4993 scope.go:117] "RemoveContainer" containerID="7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.060919 4993 scope.go:117] "RemoveContainer" containerID="f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.086399 4993 scope.go:117] "RemoveContainer" containerID="460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc" Jan 29 13:02:09 crc kubenswrapper[4993]: E0129 13:02:09.087457 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc\": container with ID starting with 460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc not found: ID does not exist" containerID="460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.087516 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc"} err="failed to get container status \"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc\": rpc error: code = NotFound desc = could not find container \"460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc\": container with ID starting with 460660d40e8b5a08ef93df6c8d319afc998024e7ab8e37faba5bcf35cbca9ccc not found: ID does not exist" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.087549 4993 scope.go:117] "RemoveContainer" containerID="7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29" Jan 29 13:02:09 crc kubenswrapper[4993]: E0129 13:02:09.087895 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29\": container with ID starting with 7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29 not found: ID does not exist" containerID="7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.087933 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29"} err="failed to get container status \"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29\": rpc error: code = NotFound desc = could not find container \"7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29\": container with ID starting with 7f7060dfb5b8dd3f61325d62a5738acd5b35b7615b8249f022413e1d68ccea29 not found: ID does not exist" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.087957 4993 scope.go:117] "RemoveContainer" containerID="f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b" Jan 29 13:02:09 crc kubenswrapper[4993]: E0129 13:02:09.088668 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b\": container with ID starting with f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b not found: ID does not exist" containerID="f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.088730 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b"} err="failed to get container status \"f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b\": rpc error: code = NotFound desc = could not find container \"f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b\": container with ID starting with f40222457182fef451d97dd1ba43c0b026bf78ad81b4c98fe09809ed00fe379b not found: ID does not exist" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.650894 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9176028a-a581-4e83-823b-a519d2907240" (UID: "9176028a-a581-4e83-823b-a519d2907240"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.664538 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9176028a-a581-4e83-823b-a519d2907240-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.952615 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:02:09 crc kubenswrapper[4993]: I0129 13:02:09.969708 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t82rw"] Jan 29 13:02:11 crc kubenswrapper[4993]: I0129 13:02:11.208073 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9176028a-a581-4e83-823b-a519d2907240" path="/var/lib/kubelet/pods/9176028a-a581-4e83-823b-a519d2907240/volumes" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.491725 4993 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:02:49 crc kubenswrapper[4993]: E0129 13:02:49.493065 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="registry-server" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.493101 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="registry-server" Jan 29 13:02:49 crc kubenswrapper[4993]: E0129 13:02:49.493126 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="extract-utilities" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.493143 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="extract-utilities" Jan 29 13:02:49 crc kubenswrapper[4993]: E0129 13:02:49.493172 4993 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="extract-content" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.493767 4993 state_mem.go:107] "Deleted CPUSet assignment" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="extract-content" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.494228 4993 memory_manager.go:354] "RemoveStaleState removing state" podUID="9176028a-a581-4e83-823b-a519d2907240" containerName="registry-server" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.496484 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.516717 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.550603 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.550662 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mp52\" (UniqueName: \"kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.550693 4993 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.652296 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.652378 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mp52\" (UniqueName: \"kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.652422 4993 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.652886 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.652926 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.672384 4993 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mp52\" (UniqueName: \"kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52\") pod \"redhat-marketplace-c8stx\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:49 crc kubenswrapper[4993]: I0129 13:02:49.826376 4993 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:50 crc kubenswrapper[4993]: I0129 13:02:50.260918 4993 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:02:50 crc kubenswrapper[4993]: I0129 13:02:50.330893 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerStarted","Data":"be420a88d18a09509831b4f2e99d2e000bd529e53f656b1061eea12b354dd2a7"} Jan 29 13:02:51 crc kubenswrapper[4993]: I0129 13:02:51.344465 4993 generic.go:334] "Generic (PLEG): container finished" podID="f730e071-2ea1-438a-8221-baf5c3ce05f3" containerID="d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61" exitCode=0 Jan 29 13:02:51 crc kubenswrapper[4993]: I0129 13:02:51.344569 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerDied","Data":"d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61"} Jan 29 13:02:53 crc kubenswrapper[4993]: I0129 13:02:53.360135 4993 generic.go:334] "Generic (PLEG): container finished" podID="f730e071-2ea1-438a-8221-baf5c3ce05f3" containerID="a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878" exitCode=0 Jan 29 13:02:53 crc kubenswrapper[4993]: I0129 13:02:53.360223 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerDied","Data":"a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878"} Jan 29 13:02:55 crc kubenswrapper[4993]: I0129 13:02:55.376298 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerStarted","Data":"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9"} Jan 29 13:02:55 crc kubenswrapper[4993]: I0129 13:02:55.404474 4993 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c8stx" podStartSLOduration=3.310776141 podStartE2EDuration="6.404449879s" podCreationTimestamp="2026-01-29 13:02:49 +0000 UTC" firstStartedPulling="2026-01-29 13:02:51.346276024 +0000 UTC m=+3455.361406160" lastFinishedPulling="2026-01-29 13:02:54.439949772 +0000 UTC m=+3458.455079898" observedRunningTime="2026-01-29 13:02:55.397994752 +0000 UTC m=+3459.413124878" watchObservedRunningTime="2026-01-29 13:02:55.404449879 +0000 UTC m=+3459.419580015" Jan 29 13:02:59 crc kubenswrapper[4993]: I0129 13:02:59.827074 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:59 crc kubenswrapper[4993]: I0129 13:02:59.827444 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:02:59 crc kubenswrapper[4993]: I0129 13:02:59.868586 4993 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:03:00 crc kubenswrapper[4993]: I0129 13:03:00.453577 4993 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:03:00 crc kubenswrapper[4993]: I0129 13:03:00.494079 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.431012 4993 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c8stx" podUID="f730e071-2ea1-438a-8221-baf5c3ce05f3" containerName="registry-server" containerID="cri-o://54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9" gracePeriod=2 Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.817794 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.951293 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content\") pod \"f730e071-2ea1-438a-8221-baf5c3ce05f3\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.951397 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities\") pod \"f730e071-2ea1-438a-8221-baf5c3ce05f3\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.951437 4993 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mp52\" (UniqueName: \"kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52\") pod \"f730e071-2ea1-438a-8221-baf5c3ce05f3\" (UID: \"f730e071-2ea1-438a-8221-baf5c3ce05f3\") " Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.952449 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities" (OuterVolumeSpecName: "utilities") pod "f730e071-2ea1-438a-8221-baf5c3ce05f3" (UID: "f730e071-2ea1-438a-8221-baf5c3ce05f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.957976 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52" (OuterVolumeSpecName: "kube-api-access-2mp52") pod "f730e071-2ea1-438a-8221-baf5c3ce05f3" (UID: "f730e071-2ea1-438a-8221-baf5c3ce05f3"). InnerVolumeSpecName "kube-api-access-2mp52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 13:03:02 crc kubenswrapper[4993]: I0129 13:03:02.983830 4993 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f730e071-2ea1-438a-8221-baf5c3ce05f3" (UID: "f730e071-2ea1-438a-8221-baf5c3ce05f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.053556 4993 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.053630 4993 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f730e071-2ea1-438a-8221-baf5c3ce05f3-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.053647 4993 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mp52\" (UniqueName: \"kubernetes.io/projected/f730e071-2ea1-438a-8221-baf5c3ce05f3-kube-api-access-2mp52\") on node \"crc\" DevicePath \"\"" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.448713 4993 generic.go:334] "Generic (PLEG): container finished" podID="f730e071-2ea1-438a-8221-baf5c3ce05f3" containerID="54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9" exitCode=0 Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.448780 4993 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8stx" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.448800 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerDied","Data":"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9"} Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.448945 4993 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8stx" event={"ID":"f730e071-2ea1-438a-8221-baf5c3ce05f3","Type":"ContainerDied","Data":"be420a88d18a09509831b4f2e99d2e000bd529e53f656b1061eea12b354dd2a7"} Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.448981 4993 scope.go:117] "RemoveContainer" containerID="54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.480752 4993 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.486110 4993 scope.go:117] "RemoveContainer" containerID="a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.492318 4993 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8stx"] Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.509876 4993 scope.go:117] "RemoveContainer" containerID="d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.530949 4993 scope.go:117] "RemoveContainer" containerID="54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9" Jan 29 13:03:03 crc kubenswrapper[4993]: E0129 13:03:03.531454 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9\": container with ID starting with 54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9 not found: ID does not exist" containerID="54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.531581 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9"} err="failed to get container status \"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9\": rpc error: code = NotFound desc = could not find container \"54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9\": container with ID starting with 54d883666e12de6e55b685d25bc01b1dc1c9e26e00560888fa6070edde1da4c9 not found: ID does not exist" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.531638 4993 scope.go:117] "RemoveContainer" containerID="a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878" Jan 29 13:03:03 crc kubenswrapper[4993]: E0129 13:03:03.532118 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878\": container with ID starting with a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878 not found: ID does not exist" containerID="a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.532157 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878"} err="failed to get container status \"a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878\": rpc error: code = NotFound desc = could not find container \"a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878\": container with ID starting with a9ff4c6ee805ec38ff99a9fb330bd60d667cd0c0238f3fb274fd5745d2c8c878 not found: ID does not exist" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.532202 4993 scope.go:117] "RemoveContainer" containerID="d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61" Jan 29 13:03:03 crc kubenswrapper[4993]: E0129 13:03:03.532532 4993 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61\": container with ID starting with d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61 not found: ID does not exist" containerID="d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61" Jan 29 13:03:03 crc kubenswrapper[4993]: I0129 13:03:03.532582 4993 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61"} err="failed to get container status \"d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61\": rpc error: code = NotFound desc = could not find container \"d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61\": container with ID starting with d941b9964a288d697c93b65b9193724b9f6952cb48d55bd1d47bfa4be7f94e61 not found: ID does not exist" Jan 29 13:03:05 crc kubenswrapper[4993]: I0129 13:03:05.200379 4993 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f730e071-2ea1-438a-8221-baf5c3ce05f3" path="/var/lib/kubelet/pods/f730e071-2ea1-438a-8221-baf5c3ce05f3/volumes" Jan 29 13:03:48 crc kubenswrapper[4993]: I0129 13:03:48.155651 4993 patch_prober.go:28] interesting pod/machine-config-daemon-bskq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 13:03:48 crc kubenswrapper[4993]: I0129 13:03:48.156227 4993 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bskq4" podUID="b1df6750-623d-4dda-83cd-dec067f8fc1a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515136655301024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015136655301017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015136646002016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015136646002015460 5ustar corecore